var/home/core/zuul-output/0000755000175000017500000000000015134131661014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015134142271015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000321762415134142232020262 0ustar corecorepikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs$r.k9GfhyBJ~F{b}Wߟ/v]qo_fZsZ-j-bֳ<ξ>|Ƭ>Uח۬eyϫ7N۫㻯7bz1[/y}U~(+2'rs\mw6鮾f?&~|3_L2_f_ṴHJ2E$(Ͼw7 +]W7t;[V$+Wxi2?<{9<;>'m_VͬkmVN:`SӸUKҳt17ä$ ֈm maUNvة|'~[.Vp|A*Z*}QJ0SqAYE0i5P-$̿b;,ľOv%\ MޠPBUB1ULo(}?=ZkLd. yK>"dgӦ{ qke5@eTR BgT9(TڢKBEV*DDQ$3gFfThmIj/ʄKJ=5OͩLH/:;ߡՖQʡCOx]*9񣆍{v3F|lQ{jv ]t 9I`V)FRI)X)yCBoX^P\Ja 79clw/H tBFKskޒ1,%$BվCh,xɦSwPmϔKi0>,A==lM9Ɍp6^Ws5!90n݌ mr"/QI&dLp4+CN(44iVz- 1 EaE nQ Ӌ[kcih>F Lbِp!\mT\=e~eADG.b~?|ђP? -8%JNIt"`HP!]V 尛a۳Ө⸇N~EQ{}ɛr9lwC Ӹ!uWȳ)gjw&+uߗt*:͵UMQrN@fYW9,,&h' UCsK٪L.2teB ˛"ո{GSi`du듎q+;C'16FgVlWaaB)"F,u@30YQg˾{_CY.hǑ sV.TD=VAKNl4Kš4GScѦa0 J ()w 7ʞXlýcsT 'zXm|oI7cGԓL!$<[yҘJqܚ%G)|A{+V#dFjVh#um`;Duhc=JS\kkZAY`] s Exb 5%woR[oI]${&L8<=#0yaKL: JJl rw#H+B|ɧJiM c/;m#NS80J=l}._?M<lu Y> XH\z:dHElL(uHR0i#q%]K&>.~y7/(.Xݚi;n%PAU?/(z_n}9U}'/osVu[H<9˷dDLF amKGm+`ùXz!"ܚ: hL\1u޷P_U{GwNNXYt\oq_@gOV ]cӰJ:^q';{wV=mdZB4']a.QO:#'6RE'E3 */HAYk|z|ؾPQgOiMcĚ$H4x>bl=pd9YfAMpIrv̡}XIթJ:|nl^/GSZ;m#Nvj{,4xPA*1bv^JLj&DY3#-1*I+g8a@(*%kޏ=S-ݑ2ƹڞ7կZa{0dqȾ3̗w.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ׽-t?}d,ZByX/&Ksg3["66hŢFD&iQCFd4%h=RA?ed}ĥ6ui ; dFF끄߽c1e-7C5_7XѪ;Ʃ2tStΆ,~Lp`-;uIBqBVlU_~F_v+ERz#{)@o\!@q['&&$"THl#d0 %L+`8zOҚƞ`wF~dWCg5o|ΔC1s`u.EkB6ga׬9J2{?+Fi7Z(ZN~;MM/yu2CݬP]Wtɖ9F.[-2, 5 CۈP$0Zg=+DJ%D  *NpJ֊iTn!tT̅Rhɇ ќuޏ¢6}#LpFD58LQ Lf~/EOF p hs[n >.]ԾɔܨH" nyiPEQ]% Nl%W+'h=TԫeVިO? )-1 8/%\hC(:=4< ,RmDRWfRoUJy ŗ-ܲ(4k%ה_' c9*%WyΈ W\Of[=?+ednU$YD',jߎW&7DXǜߍG`DbE#0Y4&|޻xѷ\;_Z^s=M\&+1gWo'Y;l>V ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~w w?38v?Lsb s "NDr3\{J KP/ߢ/emPW֦?>Y5p&nr0:9%Ws$Wc0FS=>Qp:!DE5^9-0 R2ڲ]ew۵jI\'iħ1 {\FPG"$$ {+!˨?EP' =@~edF \IG,CNˆ-L{L #cNqgVR2r뭲⭊ڰ08uirP qNUӛ<|߈$m뫷dùB Z^-_dsz=F8jH˽&DUh+9k̈́W^̤F˖.kL5̻wS"!5<@&] WE\wMc%={_bD&k 5:lb69OBCC*F!6~ö9M?^\r\ߺnqZV@z%=\#|-3ڝa$ΫM|-LsXY r# v&讳YE 6X̀v"@L'aE p6mD[%ZZv'#NC5CL]5ݶI5XK.N)Q!>zt?zpPC ¶.vBTcm"Bsp rjWhUuta^xN@˭d- T5 $4ذufw}}1L @5D] }Oõϸ~rj uw\h~M il[ 2pCaOok.X0C?~[:^PrKo ^ƒA"ZF[Bt9 @bekۜ)߄ PQY4 zF u } hsߺi!4ELy!uG7V]-؆p Qo^Cr6q,"u%neDdF O>y_:,eX?X맻c5ޖSweيO4L)69 War)|VT;Cq%KK-*i ѩQٰ`DݎGu( 꿢\cXn }7Ҫa n[{Y bcWa?\34 P U!7 * kTuwmUr%ԀjƮĀdU#ۈӕ3ΊeBO`^}ܖj49lnAvoI "'#|kf1:X l]ABC {kanW{ 6%#mϸ.6pȡstuf%Plx3E#z_6ͪ)Bppًu_wm/0}T>CUX\!xl=ZVM\aٟ6h㗶E۶{O#X26.Fٱq1M k'JE%"2.*""]8yܑ4> >X1 smD) ̙T٩XfnwFg㧤[Lo)[fLPBRB+x7{ ͽxR?PQfem9/dejOg7eY~X/Ʈ$IY&\dЕޝ{:MC@o].`c3c!I|pFEQG9$Z'&\tw$ڨAM-ݟ87I,[[(=.+>n` *3UP0Sp8:>m(Zx ,c|!0߽a`W%ATevoYFF"4En.O@`<ۃ78ϵq\FOXƀf qbTLhlw?8p@/u7ި!:E#s:ic.XC^wT/]n2'>^&pnapckL>2QQWoQB_dWoc0a#K1P,8SP)\wEZ(VҠQBT^e^0F;)CtT+{`Bh"% !.bBQPnT4ƈRa[F=3}+BVE~8R{3,>0|:,5j358W]>!Q1"6oT[ҟT;725Xa+wqlR)<#!9!籈K*:!@NI^S"H=ofLx _lp ꖚӜ3C 4dM @x>ۙZh _uoֺip&1ڙʪ4\RF_04H8@>>fXmpLJ5jRS}D ?U4x[c) ,`̔Dvckk5Ťã0le۞]o~oW>(91ݧ$uxp/Cq6Un9%ZxðvGL qG $ X:w06 E=oWlzN7st˪C:?*|kިfc]| &ب^[%F%LI<0(씖;4A\`TQ.b0.kdt羸rHl/o3P τP|ק_;%X6Q@d 8&a)a.#ۿD> vfA{$g ăyd) SK?ɧonxlM?~n Θ5 ӂxzPMcVQ@ӤomY42nrQ\'"P؝J7g+#!k{paqTԫ?~VX–8&w@a`@/t[Edso\wz|In;=&'v]gخO)0{ zz2 堛 ]iSCQ&s~In/SZ % 'I Ƿ$#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%mgo 恤hˍJ_SgskwI\t`ﶘ080ƱQŀllKX@116fqo>NrU Ѣ9*|ãeeH7.z!<7zG4p9tV|̢T`˖E ;;,tTaIUle*$!>*mBA2,gJIn_kSz)JC]?X(OPJS3.}clݨ{e!MB,cB߮4af祋,1/_xq=fBRO0P'֫-kbM6Apw,GO2}MGK'#+սE^dˋf6Y bQEz}eҏnr_ ^O^W zw~Ȳ=sXअy{E|\΋"?|NKfֱn !-p^:ZYUv`Ƌ-v|u>r,8.7uO`c Nc0%Ն R C%_ EV a"҅4 |T!DdǍ- .™5,V:;[g./0 +v䤗dWF >:֓[@ QPltsHtQ$J==O!;*>ohǖVa[|E7e0ϕ9Uyzg%pg/cc6RS`HFLЩ LkJu\!`0);Sak$Vfp~C%YdE6c>1ƕ (0W4Q>@>lWN"^ X5G-nm.8B>NOI[31,j2 Ce |M>8l WIf|\q4|UkC.gr`˱Lϰ} xr.~l-ɩu_Drd31V_ѺUib0/ %IYhq ҕ  O UA!wY~ -`%Űb`\mS38W1`vOF7/.C!Pu&Jm l?Q>}O+D7 P=x@`0ʿ26a>d Bqε^a'NԋsI`Yu.7v$Rt)Ag:ݙyX|HkX cU82IP qgzkX=>׻K߉J%E92' ]qҙ%rXgs+"sc9| ]>T]"JرWBΌ-zJS-~y30G@U#=h7) ^EUB Q:>9W΀çM{?`c`uRljצXr:l`T~IQg\Ѝpgu#QH! ,/3`~eB|C1Yg~ؼ/5I7w9I}qww}U~7뭱ԏ,}e7]ukDn`jSlQ7DžHa/EU^IpYWW兹Q7WyTz|nˇ _qˍ[!;n ^b k[);ng]ȶM_u)O_xV hx h[K2kـ`b duhq[..cS'5YO@˒ӓdcY'HAKq^$8`b $1r Qz?ۧ1ZM/G+qYcYl YhD$kt_TId E$dS:֢̆ ?GЅ'JƖ'ZXO݇'kJՂU086\h%1GK(Yn% ']Q; Gd:!gI-XEmkF}:~0}4t3Qf5xd\hEB-} |q*ȃThLj'sQ %؇Gk`F;Sl\h)5؈x2Ld="KԦ:EVewN ًS9d#$*u>>I#lX9vW !&H2kVyKZt<cm^] bCD6b&>9VE7e4p +{&g߷2KY,`Wf1_ܑMYٚ'`ySc4ΔV`nI+ƳC6;җ2ct"*5S}t)eNqǪP@o`co ˎ<عLۀG\ 7۶+q|YRiĹ zm/bcK3;=,7}RqT vvFI O0]&5uKMf#pDTk6yi*cem:y0W|1u CWL;oG^\ X5.aRߦ[_Vs? Ž^A12JQ̛XL:OEUپOY>WK-uP0\8"M: /P4Qz~j3 .-8NJ|!N9/|a|>lX9T ҇t~T1=UF"t; 8-1I|2L+)WȱL˿ˍ-038D*0-)ZyT13`tTnm|Yhi+lQ&Z!֨řoҒ"HKX 6„=z{Ҍ5+P1;ڇ6UNE@Uo/>8.fgW]kY0Cgcu6/!_Ɩ} ' Ў3)X<seWfSv!ؒRKfs%(1Lhrٵ L.] s?I,HCԢ[b C-lLG+@_$c%* _jR|\:dc5u= A@kUc\ǔz;M>dUN/aFRĦ@x؂ǀ$6%}N^ \mQ!%8j0dUo=rh>*YȴU3Q,̸*E%59sTzɟڮ2kg ۱wEUD3uKrr&"B:p`\E)j<).R&#ÃecE,dp"nPS 44 Q8ZƈKnnJei+^z '3JDbSK;*uБ:hF ѹ @˿ޗ~7g9| hLXULi7.1-Qk%ƩJ4^=ple;u.6vQe UZAl *^Vif]>HUd6ƕ̽=T/se+ϙK$S`hnOcE(Tcr!:8UL | 8 !t Q7jk=nn7J0ܽ0{GGL'_So^ʮL_'s%eU+U+ȳlX6}i@djӃfb -u-w~ r}plK;ֽ=nlmuo[`wdй d:[mS%uTڪ?>={2])|Ը>U{s]^l`+ ja^9c5~nZjA|ЩJs Va[~ۗ#rri# zLdMl?6o AMҪ1Ez&I2Wwߎ|7.sW\zk﯊溺^TW^T\*6eqr/^T77WNZ7F_}-򲺺VWQ77V\_v?9?"Th $LqQjiXMlk1=VzpO֠24hf 1hi D{q:v%̈#v^nBi~MefZF >:/?Ac 1M'I`22؆DT!/j璓P åiw@wgRCsT~$U>ceއE)BI>UljO|Ty$ŋrwOtZ7$ "i 8U 7bSem'k?I+/SczF_/k ``'1Nxg`$׼ٞ緪Iɲce3aL]_UWWץYI=Jb2"dK][ զ(eJR,=ie~_ՠ:0~_u BT"Jj UY%}+mߔc i쫢"D{k=*Fëfsܚʺ'<(_jsDUa/Dȷ${k)*2nV3ľUX]{>U9)ob3ym7ף Az /|W".x{L?1- `Zf~ ! $i\'qZ\9ytJʰxJVtJVa~lqn|o<7|<keE#ŕc yqLÎ\ݝpC/x^\Y!2Wl uB+bj1qd<Ӥ5i~_qɓF>heQ53Og~Az/B״~$_yJeٞS!VC)x`qÓ, ֿ;<&"3ҽ:v}8B7gsM?%Nψ#' "9'vKUϘ 9}| q(I_УHz<:5;Mc\׸ξaȂdr5]^jx7i!/ 9o?M\7M\gn9SK7<5A_UׂV{,=PF ƀi\o+d2r }zͳtO|mVcC`]?Mf븎K$߈ ݕ^97KlI(! |fVLf7ݫdnp9y₷Œ׋$nf,6CLIai2a.X3!&)뗳;a=kx}Q*Zk]jWT hKToxVF~etV.D%Vt$6E^|Nnf_}>{ٛ*蓌z}Ύ;4gQm'E_'k K}ϬǾ_M¨v;s>͌Xv8yvmǛgIA+'<ϳaTZ-ܛA*~aICBu[-G (|B%hx Aʆʶ^mh@v_SYL>%UŽBlྂp`>9"a DƷV:Ea:NedI2v?v_Tpy;כ]j߃=H)x։xpQpg?ue4{8|DBݍL5qSN A/??QrSxSmur49h뻛wK~M<L`!] "7*.y%J!H6]|߈pИQ,Au[A> %ǧ~xl ]A*nٶA0.F5 mgo4, $D_VHPHQ}mQ׊RiѼI-u>: #Nn!3_p6Syw*Ţn^OSr`I\s8HZetEO[4c Ÿ甆-r^֋6|JDttg*ŋ]E&mv.&y.v?`*/2Q85'Dd*OW/[ˈyj|^#2w/'ƣݜ}*w@JKim^XEҰI__tn^Y> xxQܾR,˴Χeݍ-W4ɒЉpi"XAӚ*wЂLIHK#ZN_H{ >-!EEaix,2iI$bʪg`fx;U`Mn /z'D#;RDj۳G?%H76ItL%aE.<|ŒH) Y)C9tCu<Ƶ_ZO qRD2&j"aWM8vm. 8$kO2_,-Pv,m(nSq!54lhP5 B$v> -8iAu ~.=vr\Tre-+D:hXacnCk )|J6Gy6TSwEh@5*M 1|R{^ٿDV]{[}; )O22ޖtKКuҽNL13};<{Yf"eՖJ p9|Be<ݴZ2ϋJ2 0:G$`'$s@KcV UYR/ylɧ+ܾ$aIa2ZOiiq%"G$ PO L[m-.(ѳKP<'j KIDjkYR-*V lZM 0]Y3fLۣT lTKq[NȂeͣrR_ sæU%e'-\WycTdzȖ}{P[$,`.!$ҜZ]5zI1*cW>2sIukANdp- \`6 ̖y4jAq-q ,%ڪ)U< zoLU\WV)L3CUX7Q늨T}}< ²U.H).6 2aI,Fkjlc{Z+IGEPjl_MgKn hDKM +@ΠRΕFG7q &w",m_(TVdEuMfl2m}x).SUY!bM:uŞj7ɦqUIT^TLZRxkcϓZQ{Z-J!W=ݶx#6]KjvT9"+R_zlW"O7V!d*\!* BDm]%ݍqCq$s 8\2%.eK$橈eK$ZؒXѓ9JrEe>ѽC)>CsC42}ѥ"12CO8V98|V=qAjDΓ݌YtOP,*XiiB 6&7D=SnkL%[ԕB0g s !7&^]Jyx+GBhT5L׶Vn&pM]^r!2QTy)>I^n|AO6L5VoҺ{{ʲ/%ؾFDzFg,5IWZ?\ʡ*]ʮ5v!0YMQi12[zV=* :,:-I#)_I`p*rH Ͱ5V["6{+l|~ߒ0@Ęn>q~F[1⏦c<ö`(41KwVbeJ`qGJ:f{t`w2xci͘=fIN9#N5+yK35#u PL`0h!I荢8h1 rAnbc^ö(h\o,ј#o1 >AhBHICA*["O@H%/8v_>z ƌDq$Hw4s(k?Wd+@=#x-*_-s4-'nsɬ ̥j'W74(_68N4 /Y2[4B`׍OHS'#j:NJY} ..a=o!DzrI<7|}C8Vny_?]v;¾Uä\_ӄ^cؙŀ.CA+y10ˮm Kܫ-iPv_wGC|q.WQe6(:urIγ<6>G0:LwY.B1u*Al_Mg*X~mJS,t{G~w+b:.e>ꪣI\~%AEqCXhډ" C58l` {p\ci$,ݛ ٻsãW?(_U Jgي %n;'Ȟu7Rړc %ؾ\,  78rh$_JA/Wzq#x0TI(?^X]HA;= @NT TB hDd]]*I,: p@QiK؏=xO@%)I_ yE%S"r'<"Fb { 7 0šHP~nOMWD(]:H({N~D#lggH B51 v{)=˵!A=a ]eg0D/".IVN9%?S D]&%Gqo6``+xpW {`!pߔX};Ja/,RHZɔrzwvx7ہZ^e-AH]_̙Dv|62 RgʭIAd6$K~0ٮoO~&]: ?4[}:w{9 tW' rvw" ]=+0pxÞIidž ayL00|h:C G{cC0Hpձ/ӇY;#΃D[IP^o\CIzM$ Nz= #G/dzyE xof+ E#C{tZDa:N#<4RRgBô n9]°7)03G w\ nܢTXP~u0Qf$X% ]$"e=H^b>ŢZ@xt,I"@f_>Y%&NT$\-52Y<b]%u!uHpssvKq^qYL5$ TN$D:gSk`WSF7 0;KyWCnuu^MT@as,Y= ; nÑApڂ'f6Cy͈ς721ܷ)7E ~xbE\g3nâ܁B`w%uƅ3IL"d#UE<\s j6pC @nl&&Y=h-JI0X4lmo&f?]#UhR, IsRT{. &.p]\e %I٬8k`#\2(b19pgP]G:yYHzc~|ъH Y$sB>`p8F6E .*K-2).+*?hs<ǍVEGg䚑Rc]ץ;v1ƴ;6?:}jBeNRy'dI'YvvR! V]IݣX-& @5ɇ/AMG-@SJNۮr{ľXϸD4^[o3f̦TJJW:4܌@@IG-|2"]]>Gɇ_!F_ XQJ^%|?AZ+y4XL㊌("=h"i Xo Uƺ!i'p77jޝVܦ/wZ6q"ATCJV6FzS`>@TA)!ʰ,I|kk/,t#3ʄGܑznea’ɼ&X,i#gĹ$I|w~&. -.í$O/'dzq}U%Mp8aw[l><-nڟ^ev~o_/-"ePy#FA%\z墔Rc?]&6L/f 1a^ygG0 ֣ 5/3(~?fW#k[Wl\kXa3 @Hp ?U5wAPryv.C*͹5ItoN eb u.5[7OJjMz_qqՙxC9vi?y7G2߇}ȣwy c{`6j{}C}9f -O⹔5WF0d5r7JDdn ǗaqU.͊6ĻI(dsξ׃j7;f}E,=(_\5t/1E{ٟEeOk 87 Zq])IQa&G"aˌD0VX6:SwY[ $1]gA[MYTvmE=rQofhKr}%m A;ඊTT( yYƠ=oM-lPlwBe$BƐ]Ɣ=bPG=vXN9;ccet:|bI)߁PgPgB uFHBu uw ݝPi$݁PoPoB FHB w ߝPi$߁``B FhHB넊*Fx$bBuB w'4|# K C)K )h󬸙Bohb!쎅bV\L|w #C!$$hfJ唃K{ȶZw?0w@myl^/¿]sRfwUS xA^!22s>[~GwebtŠQ׳)yӠl1?O'Gg'FGE>i } r,2E2BU.?8'.RyZ:Y&h}<5N0r\}mǨrZhi"y^s֫ٸ( ǡbE y2wkEdVHKS:U6xbfXxx$wSii@'leVW{%&PXRjYUgxkˬF.Bo8a^ uβdSE/48s(u?`OK23F@B~+25>`k3*Bre4C@oɔ|U1Ne@̥lZ g KL2D7 U-tR/̚GEz4o@[^zw0!vU[_oKɳZ@5sݿ9B"挊"AdH1)|lutLS!s3r-C煻٫r/Nj P^108[|~e^줪Xdt:(["*Rh\4G9}"iе7jTe7_ <RLSJ{ϵ kCyIu;KV!?48dTy%sÒO% ZFl"ÚkӢ@ Vx.' 2)PeJ,¡:bzd\N؅$a0!֬2p8\zNA/qZJ8v 6pAy;3eH4Gz e%L iQ>Ӣ|~W!Yf㉔E}"sz]Y sTS1nUXy܃ ~@ 7ق>R@ v o oxS;РȅAR ۪P,4fTX^)ϗ#M>8r!zL{.P=*weܹ . n}&vXY%k)RҕY;OM4X>[*@}NnPXK" B@Z L(eE.k$3@ԉ ToNGhIפa",j3X3 {@(/Be 8LRgyH[zmXB}jUj\T:EG9ND0D&͉1\~E[;%lv!Jd}c9P~˙BIݠ1 nZÌp2r>U' m峻Y6(|&CPfm4W-/%̦Sr从'1kB,`g4߰UFޙe޷75js|f(Gs>]]ehyITLęƙtMJyǬ4Qz.Uz)Rhno*ZZ&XpLיptvlk_؁Y'a48,Nd۠ETɩNՒ$4V|sׯ,&/iʃg2Je=1`ִiC7Y|aTBfqj Rӊ= Ih+yRPԠlpkd.؊_j (4R+Tcu0ΆhJ0օRI˂/^$BVhc\R$AY=A3j(~ԍ>FҨUs)\fv)ޞ.NJx#՗R 08LMfYDSYմHtT]S|V&@~{Қ#ÛrȦsEYi-Q+38H6)NS:vNۡE2`QJ|E!*Rڰ;)άu f,9|I`JeMKȫ E -czyȕ; EQZ]JU':sFitAwOPê[-c7֝"9scm'r*bc؏O7xl>ԜR)_.V3C85 az n5̗c4)1k>:bp9 GӃI&I>! Qb@oG&$ŘP_q_K[w^ʓшD%`!7 2n!ER,!7c U#WySt GWĤ%BS$AlC5&t$a:p6"I,*O"L#>{(0XX ૮ʄP"!f n)"洁m_szEm18/-%ګ*|?.)9dViǐ"._cHv;8.-[ȡ?giY6H&b%J Z8^%'P"bcXnڪ2$9"UC%y€Li(29ƒt1,)-y]y a`M;@e}k Y<-HS^}F+H5VlỸWcd4Kt}Nwьu 38vkLCxdۮgSaq\0Z%x8] EFb1hRalz!}7ߖ4qorި<'V5]޽tq0Ç X1lHt[U7]pLo\Hnm)zBi5_ '5GBk%Igm$WR31SÓIZ)`2v1?XӮͮ,^q? `xc4pYmsU*GDa5x2g^^.D)h EbYRGpm%H@KMfgLI~Z "NH=PԌϪ}l"}_:[[lfrH"*_+$ĤЬFoFFHWzIgE4|]HD÷VtF2pj,q#ܩ/ ̪L(ޞO}ѳ[H*]hl/h/n'7t- Uœ}. lca%֫xoTa.7IV Uxl=}WCʫ $0jHä257o\S;Ґ>.FĈ8ʅB C 5kU1Uƒ9H݉nsTpB^.f p*WŴ:kN,Q$#ԡ85Rh-[8V$8Mg1\W'5,yzH X9kfZz"iDz^sXAd3-?x̥ u>?:pazZ$;@7ᤓm }yϮCa~uCp^II;G! )Cf4dZ~}tY=&N4A=.7lp)*D(V r8usAyE84x"KGM֔qg5}*6T=vQ%#?vꋸYe'q)뷗Weu nQIl0yRDOXoKM3ueh,b,@c,jw"w3!~V8Du6˰vlV;!C[ZXA`0,˭v1eXG<IVtQDjld.el[nW_LͼjI*v/^Ү-|:!u +A^no5)K R FD*mXR$whS|`n='z/լ&R6RQNi+*?XJIM)w!ExGWN+|< -0lJy"?=y`DtG/4* Lۗ>fzp&\e$zx÷|\/oRkWbڎgw}m;<`]@nVMh҇bx3PK}|,}6} 7},d:ARjRb}ot~IgF]]p.U+Y0pQ0A"ϊיTDnpWLK|9(@Ik!3H&k""d{cnpYEmmی jۯbq R18 q4cqpk<VA-Ws,q E%>uqKq<( nݗ>?ۖAhu MK<]&:#G\54冮v8TqBR3ELG^qv4ηWY~{5̤o= WۗӷI9wĎt>DlKUFhYzzPRf)0YmIgNw=$,MyľXBVQ_PW/Rel Ԩi㴁MxK>P[h9EvuPtwiMI[t|?Lݡv0>vwVvejYlmf{сkI-ْ:DHLxMnAl! `x4ۃjxiߍEcƟ1.JS)#8e+_zRV>X8fqB ;oEC|:OVBښگsКrǩ1$54)_~-_ntgW>W^0 'گ0=\AD;+"94ZIp,Që u.!jM:%Cntjw#>DhN& Oe<@a:ǭ7,E")XD"E"8`fhbYהi|q^f)О$Q~]{-cnDZSq}[._S@Lqg >qŕc/KsNHvo_7%P\pW%8*]0kB蕝8-TV3 vG+oKtb0~^m gu9^GG1x$=9ve[0FQoD?vbx woԷ?Fv`dC{aMPDbʧoՅW%:`ڣ*N}3P\޽4wٻ~ ^>Rmhн9U׾~/BMec'=vGtCwUB'W>~Y6y5PVj3;,]Q믵qf{}o\Thdv+b߫8 @hl: K&T*Sܺxlҁ;k}*~[\9z+DmW~^q8S^hoSX}0}W+Ȣ>#[,ānab{W,S'J`2V sf&_`ILfkݰ)vúch? smE[W+6H^yFriR[\u E{ܬnh2z@䳰{A`ǥ$Q6P̽@q8\Pd3鉇.Yǻdk HLÝ1* 5=[c&`!MbM fv?bfm&&gloqܘ~`z[cޓSkE7-&(jldMZDY\\U;bCÀM:ddOy`~Kd273n]y#U<2K/{6B;+ڠ}y%S w+O&*h\G?ZԻW>X@u+NJ=)oMk))h {ofj"Mش6+yVZ[GZ7L+fa]0*R+bWM1[=G/_XPnO Q[l 6;tyl&`T QDU3| U i٘[t1遴d| &/f)M7}]wck ޓ>fw3] z7n;|"v{(+v dp) ͏Ϣͥy3ĤY nEV)m< w-\9r&'GSY f5e!>ܘdĒ+78ݧĚ4DX8*aʟb!1+uԍo;f0X w_=m% dJ F[/Ew)6VdVo"T;9 q}%C2#oC;uWpx~8EOZ*b/&i$.Gh$9,S+p "fPYw#"_KZ!M϶biJ,1]؅+%mPFq oJ5JD?Xոɦ=O.r+_p9;d %'Z6[uδT< QmKPS.ha~,?6 Ձyk5=MBmE1 U&X(,.a;4J+J9Ff_qk=H*gP+j;9Xg\jJrL.L眒!B#ͭ PysC=39wF{\dO)CF${6@zfB(pmB0m17>J0cclRhコ,$i7m#W7).ϭp"xc+T@K4j&8ʈγ jk%L`it P 8 D;UBcp0$ G m)́YM"8xL" -SSheg$}p 4@` SI(; S#;]c-֣)dyX,89Edj89t[oI<;>Vuf6Z8i@NHbձm 0iPNĥ9F^F<YLus=JB/%2\1T S\p[0c Ԙ#+rKNJӨ91 e"p&1g$> $3ВڢBJf'kqF_!!-η k${Fgz-do[FQ5#'XbY_"NpؓDu>1*$Tr`@}xa!Я(m}K!P3)B SD@ S\TCzUr:@б'PPϓ;sr >5|6W^SnRCT!Qa,YDU׿LF Ǘw{ސPpj/M<,?ZaRJqLi$1cg$NSH|0BC2>٧!~3ʜJ4f^ ! $:!r 4sYeRd|`.FdjOdʩ~mOj6X)>>JtFdǡ1I88mFaOylJ.E `|RҠ=%iℤAӲ J!~JҠ0;iu~T'%@Fd6 ޢ~Վia?="C؏VE ,|fSOtQX3CSkUR%̻ѐfN`D!M1+-DQOO%uEt&zƝ_8 fR[K /vTx-dx vU >S"ڦZ#+aOhAFZ'wAd:Y\Ҭ ;r[R $&I$)P1+&1" =-j ͒fGK o Y.4>!`VbKdˮ- sC'd8ݑ gLN0'R:Ԛb326PN7?߈`~3rK4BS<)ixBA4hOI:ᷮ.ߪSu16o[Mb Ɵt]D#VS2e>kYcU۞ZPy_6,(!^$һ@@'w%TN_ ( ;Co#g% t3a[g R>hDB׌\|qSfx.ypm -OGdM+©>wxG +0aot|Ky\7bsu ӄLI`KX#:tѐS+U8w񉃌E6J^`OH`~p݌CmOVP%ap\!LQ69v%"7.9-J"5PM$~^iȕ԰=JzpoUR^suJU>hCxwO7=X'Ά{W{>{[0yᄏpM6_Oƫg̊P ?\*4gawE߽*w?VG,n^@%ͭs>/^6%Y'{fDѢ1 &JsM-½|}TGo>Zq1oa(A:7 a9]‰X͑ܒoVT !J\Uq%$V"umwhԙ&*N4\c.@̋ՕJ VLX~ἣ~O  $NF=o%ǻXEXZ#` 7ޒp`dqͧ-lS(ωU UoW<NϞ7} ij%5<8G䯸~7ltwM F[bR^7/pp>÷ke?>\ K 0fL3"q;ǪL:= {Nr ߒ61 {NOK}ƘS7N--aE^ٓY^z%#wg|K(O -;#=5jv.>Z1?Y|r(e6cgߣ؃Oi$$B*JJӜěܖp&9O_&AQ3š*7# &&]-g0VhPy L׌\AQצpY}VP'I5#Sy_]];4Rk`JOJJy*$"S9޸ܰÆuo e>]N?zqxmn8l5hތIT7z'7V6}4S<0G34+|_0rDmyqf%d||-ޡ7EmPph ]%,-m1.WȽ(Ef帗|wGopo͋Dяf{_x1W`/1*TaÍ.Y-vUAryH, 5~V\i9%z뇑tأ?wz׭M,f>Rx֕f?opd@U։pKaQ"w:я %bEȋ6-yF|/Vj@^}x:Ԛ{ƣ>YkTk2tû7,ɣV1OhUˮ1qVX&;џKN"sLx:l2-{gAF6n4XKesbOlQpxCgƗEg`*7w4lWB.1|=ukl[zh5m\b!nq!cUA7VIB Χ׊JY!/ozK+˝>]0]\+1{rJqsh.WBȨd%Ÿ3%xoҴ?GÊq>ql'y1ß׃" uɨWMA SW}zL!v6X퀋)8wW~:3Sg>ܗ{P $c #3f{-w CB`t,Bݵxq)i0|{1WfG@Z;"d__[z.-K{cj?j;̿FlLܟ۷Vvݞ.`#gvk9kc0L8HwU^*jԩ%+=#S^K>#39#39#39#3!vf!Kq[$>aη$HNan|AmL^Θh߇0Qٞ q;NIdZdbb^9Wq m eչjFgwP[c0e3p=m^ d|g`H'NhBzneTlA\"}p XO/ʢT;0HRIaMaR'Lc0J*[)&8qL [750WP bϮ|Ϫ ːk!(X Tf$ln"$m|5F9'c=! rƜP *Y"@wD"EJW]A"+¥{LrI5F-KPʽkA@8u9g]9rLkPbTSPB+M:Ur [A~8ܸ<)6…xfB7!YLbGzdT*5co$QDj͔JJC(jI$?@EzU Fv1Xb4o$Z$ф]batγT""LYǎHB!tB t6+Y>dꩥޙM]L.h6}nnۄdAٺuT]..yyyyA5FO6ZlϜY-bAI(3>6٪.>'(PʍxRΓ`uʰHY*fz#Ir_iؘ.bwEjTR̺uYHEaL28HE46цg Wus^qa}{,~+\[O?lscsvk!RgϡO54+:$w,Lt)L"2h.駳6ɴ(X Iql}׎?|5'J{~" {bJsY/#f"@`Df/H(7_٩c8`΍/-f1%ir\2Fpe\+}NWNQ%gnl`+ܙrp!.tfd? G~M?S2{7ucOwK\Dጯ3;D`P,+Dc0/ FX'*̿׾~\-ke)GwOyq)t-Z3DQ<"N:J`mFP/Qyt"˿0`R )xXG'3R[Olt] V-dr8s^ M$H]Nu7̺i$5 ,'mp0ƽ)Th=qSw1H`~o~hQc{Fc[j6ƈ'c:I)kn:&J\ EAy5N|*ZW:.I`bݫ Z#, 'k=3rGN`<݃'>f+, NkYmGSq. #Z~BjA)Z仦C?QSﰿ5 p jCWT (렞ay1N|"ƌ[S.vEz!qT5'%p%JydiMd.q%Y;1zXħbE5-85VIA ]IIi69ԻCh]Ttlg'fC* >p5|COZ٢G}erMI DX QYww9'6{!BG/DO F1H%&`ϋg1Cx缉'a ~ oߔWM6ZNt ur_}T~5gtP]:cM,hK8`%*#` qHH۔p0x z+wHU`._2.!iIG9Pd[TlmWv(hzЮfl`@H&Ӓl WZSr|D< ħlseԷwet7$Lid*$s69J|"zceuB2L⏅br(Q+;-geNC=PqSAc''[kh Hx/ґSC0i+cIJxWXd=qSA-}^Hb;“H=0H.#6vct ljOwnK+bJ[zy΋ըRIŴl&`w |Q_%7w?7J|*UYMSjV!bЎBWl ],Mɀ=qS!.RDCJњ C8G^ J|*̹^w-yaCgw[y ;ک|f1pP[ ipel觛ϫ (R .N[[-Eħ+znjx(p;X68-IJ LEcѭ&DP!׌̑ r/~Ħ*KQK m* DAQQ|>E(G(_f.Ix4$&.%OXk)4%aLExħZRyKWK6Sê)q3sw c;NzJ (*1 T倍{-63$һNQtDۉݴ\L8`,̶S@! +OrrAiyp4bħ%|xfthy\B1;/Xc<+BE@ODP[,j|%~h8鼭Y&/Ico1\ WKo,{ Jx(r5ޥpS^ZI8_$v1ޗFx˽~k-ԠFz1_%xg+<^Yʞ֯ݭ!~&<_?[፵\1oU;-a*C8䌗x}4C'2ʟ$G uD[xJ a!(gukc?:%>r;_hucmCnp\DTɔHX0XB$lr$6l:}6D3WLiL*9rxZj@[PL0kRM(Ʒ6%R)dejR`,u$ 0AJx(:Z'C`4 gϣ1 /:0T I~WlZsHuRU,jASZ< O1z1 29Sv0MJRF_ y*(< V1?iDY.)Q,acB5T'muʀ=qS!jnWO F-.ϛh0]/vy_m* lxA<3N&׵w| j( ~Ah7(ߢbn1+A2-z=KCb]P %>uF#D^GgL}`b<\vK=p5eKWn*1}2-Wd{IZ@| ;gˉ/oV _bl*S\LD+ $&yn]I!t.y-J|*2{aÁԎ_ _ # =v`}0˚'Askcj`mzħt_ʧ_(l:w(vOą*'k1+LШ|>m':nT-oiaF]|~anecjņ9P yL'B*Nʵ?->9[wI:LӰٯݘv{^&cCy6ZAQ={qM/SwhoU, `|JL^5p)vxQד'q;J|*WifߵGrޣ\R.fQ^OcnDvSb_MbnlW/{(# !SX\Qw3ć$@:rF7Dd$p$/z汵V_4آc:I~Q0ٵej-ӈ/y6J2櫴BK+tFzF{hcVxJ+[:4c6ZϘMbNy\6u%`u&+Yt+orQ|7{;V0z#-0_C!r= #YҌUф1b_.g NӞ}/"BG%>r;c_#naN^;))8خ`@,p]bN;}60#toaFO&&?nT‘`+W~!C.׸{Z >_,ȓ;?Pl6Z>r\7~Fˑ䢩SNT_S"x_uk cZN7ܴg'_̷6ISݲJ!!tg%e NB}BG%>rC( VWڰR;Op C֡QxvFe[uyV5#76pCKqgls˥y0M665BG߱FO"3tE'g-&K/h*ZᇳEx0ש掱~N=qS!y-fUΨ- s_E;ɫ[K{vء솯y9USuDkr_R)&wjL$P-Q^Ǽc-kxGG*S!t; І׬9ov-oti+^w0FKv~7x;5tcďMHNcl4Le /"WqDaVe'aѨ!Ɛ(+.ykc9BGrT+&V_ۘAm 2ѱ+*pq{0jBb ɻ&GrW}B ޏzp Kt(&9${fHU$D:lhgH r`.f(X?x)T:Wh帱Cc_,cvF(Cǐ ε@IG6suz/I w\x&ںG[7hxIO|rtul_crį]renJ b5Av2N3tU 9 0r>'Ыse^؄!?7q :4G2G{Zkז/xB{g6Cڙv;9 ~C,3e3+>/y9?`hQpe݃[()Bmi^݀1vnuKn0wϋ/%t0d} a6{<+,cn4">Qҧ'״9uA:xHQM3l,#"n++#5)kG) k 6=Z5T+!7/q]vib.x,4:5 .6^ acJUXgt5^}E/o^io77^ʩqNx۾ xZ5|(|<,8!Lc̥TKD1\E}X/]BbWU(S(+ҳRp9fyjau23:q:l:xA vz+|Jp=ZTTbd%^a򝯟ˋGQ~wh9W:vN2%s9?ŷ.= <:t|',}ιn5<ߏ8> /^`pTPWpT蔦EN S@StZGV~4K<:[<`Eѡ㻴,{jXDj p@%iߍfQZHBzk4J}nqzq6d=,f/~IU/M N|/:n8_8p?^fRTnPGk*M^^z9Ac*ՋP݂n6[a}:}˩~рuلϦ/:-1Yȓo&ON4CMMInއiO+h/w߂]a:-\_(6K ^u&ڄn,i>4*0kQs$n Rsƃ*VOrىQ_ABVBۿ#dm#ovbk cnQ wQNbX[4T@=5ݑޡo- mBFo6&D þP0O"y&Ɂ9-r N`|F_ "t)&BFh>co嶴(׉ߡ'mDX#w H Fu/$,_w'0]pkZySTR5Xw Hӑتź\f[ gȷw[n\zg*Wb=%#ES!Vh 崔oõd-F έ0纮ӺT#JK)rB3Ĥ RQHs})9!]P3PɪZq9.*t"&kH IQ9ayz2Fi!JCd.^70tv} j·(pq[܃׈Bm C\} Hx~{n mJo @_!6k4 s:H)) G3Vh R!L)3T֩se{#}X`E }57_-N" irE( )0 hER 7V^qAFo^ ˝sv*?O ѯ}f?U9"|{IfO۔ 6^LKߦ& ^!۴F/8#ۧ;iAVrK #6/U!%R\Fd:PU+ K+ lW 6, ;` \YEb/"4"(Y3w`f'AX*? {IRK!B^`4z42S>$b?_:S~L %)8via-bXH:5tQ''~2- .PO̧`8ql(,(|^6mʏ4}N6$j_xPyt'VugṿWNOs)PE̥viCx2a%%e%P7٭aTOCQ+\l(sTz6çg2QwVIm277Ҋ=}.aAh`5uImXt|liJD6p* .j"<Э&)IT uĝM+s$]IJ|$˙Ie {f%2f %Ҏ.Rti!3 uq8#ZKgghFDx(Ŵ9T!P%l]&T׿  X??y/V?IEJ}/a_I>K3dl4,/, J="|nT$)7xy<ǓQK ]DS8o'"+!tr+,?IDQg$fӪ(q2Nm9Uqp4 VǕ1^~_OyIiap s]xM8W&@}[uXAT. A.F|BYHQ(@\20rPVr w=dܦ "odU-UHgeI6XCt@M06eAAnY;}캡knO{UoMfqi.7 ^zB.e<2H*Ԉc]ɡz _]v[Js@DbXw8Ξ>X2w(Ε^oE7a,2e*s A:ψŎc@`ؒ}- z4 랱k,!w݋4ʥ`蓑LovDDг Qj`Ijjo!|P{[o;V4:*Bnz?E",Zh4/,n4^4F?-O|<rU8wwj+}|X'yˏe?ġ0J(N2auT}j96KȋʗճeH [Y'b} x\SZ1\&qh(:'z u<@GsCNR?QȆU%Qu2٭]z3@%?>NRKg/iF/OJ@ t7 DvTT,}:hJX^C:pd7SK f4d*>mtQKxg|+:[mHcmhmK6.66㿘SWPtmm NPzk-">>~#UF{/~ =3LFGs(+ _Ђ+BT"=ˡ^_<ίw28ul`0#~q4U96f3pmʥ. D]+Ѕclw#XzyT8 SO|s}J~D-H"7w.)A`*ᷠGm%uI`ͭ?՗_m/b`A*2ZrI[ fKz:;n1Y]`;xq\ho0#=:%-f~q;*іzKWM!Nh+0-2VKILݧrn4Aw1WJ""OJ{6AcћO5:/tc%0ZrB;!(xdƟ;cs}~ez$!̝Ov6bKl!`yc%F3($V>^\cÕc#l9'NA)9Ilxa9HE `+2\"dɸmY;`uzlᨍ4&14-^_yZ%)s+AY3=+Y0?&!=|+]׿8ǒ6>s?0v~h1PRhg?*'u(ñm["R錗h4m&H1 39"'q%_K- '_ÓB\j"\H$4Ga'C`*&g 4^XS ;ӘHGsK!n]K &)ĩӮciyau5UpQrB̧s>JI7U %S"=݆"gNXݖ\g:3'w5xhl.DV;JobԉMl~p=RXn'#7b ajLgL( J#5bB @R+\lꓻjxCH@b Qe*a [S}|b\.1/Һbqa {I},d6$tQɤ&~{v#;~Q^_ BB1uT) [BFr c%Z"I'bT%7K`C(ZW*.ْuŰxR(F9 #RmȤcd^u&:SSmV!l^QFǚ9 {5mNg,Ԑr@ \%FCZf,b )KBa"7#;Xr6_anu@; J!`IC \5?Oݟgo\ʅ7Aza'?]\6*,fX#{UӋG(~Iu#)HWnRɂwk v0:-٧F OƠD WȌ~4v*CU5eZBq}Ȟб>4o|1UgI+¦In׉b7 ƘfԨ+Y(fW!%E">>ROlF[U~]Tb Q3'i4Z WOLf>hngN)NUBM򿞈 `$e}-<1H"wg)EvERVBұ2~<wX$9)*; I)2R@c~g 5ПP}7n6H,ςɓ<67 bc1k1U?~Y{ɉ4P QDASf8ڭygE Q쨲UvrNm>=@fKZQ1)_IWbCHHb Qnڶ"VWﮠGX0Uvd$d 2Ruɬ>drvW5A% +1ee^e_'poX"'Ivc!AL &/選ttz~k-&yFy}qT2’W POR2K"`gXw{Tr;_l<= ,8")`k0M,F 6Ѹt;{vnԫa)W1ƠQ5P`-^$e47 @/,c.4#QߦN.gLCQ)-v6\ D0JҺH~f>4 !9= t{P>fH TI,^ne0pqK3e$aGLmʼnʪleǦ ?B]E. p@쁫p%Kv1٢in˭mjU66caIJ!l8`d*lrz}Q9MnV~?7ɖ) _5Oƾ#jAB~)y~.l[ZIpo|{;d%sK'Ht S׳HD1VT,oi(o͸jfW,EZ0u;F9D! 12XLJ"q2X'd㌯| ^Ta" >^6WVu|ߋCr0V8),:OpCP s;-V,7SP,NQ/m]76$Wh>?so]QU cch.+_7Pvpt}ջt<,pp5$XkXb{ee ^FI7*b:eq4zQ,bŞG/?4WX@Z9uCE YsG%wJc'Y \Q.%1`R}$hVi,lQ |շ)M}N)8^h;P8pvCxތ(7sg@"kdUǛ"UÖx.edt}G5i5+,+e9Z2k!߾!Flo#;M5Vʚɑf#M4R Su6( 4`HW :q-u,"4`:BYpdQf=RaBN=HI!~Gx6p[yu:\ ,^q:8mM޵6v#bp%;da'W[,9|^ߢ%KcGA-,;WbMujīBxm>6zv|v+]ӌ (OLs߀Y蕕9_d&cȳT!o/U6FaUԖVAp9/ A(czfl 8 ;}D&$S#FkuNTIYD~NrBgrM(7:#T9nGJq8i|jrvn]gTmޯ99q<][y]H8B rsT#-¯NJG^LK;6J< PRj, [C"s&aMDO,hn0X cеNX㾂xkwot7m3`_3yx vu01~`W㯮IpC`52HL#:ا ?~'5f%-CIW1@BNKc J1*N&r@b&ˉt[gҭO ֏]\Dmx(oKݨO-hw[f?|7tnIo?Z|wZvlg=(?3.k0\tNbR𽮸jt|;+J7:`_Z\L^+N:I_kC9G39h7WP>c]23 ʎ<yF6򰑇Z &I*Ca"$9e]`Įwt4tc4=}w&ٓ/Dc%k'ϑ V@G'o^UzOMʝJ HsD ,8<{+58|.%8tT ʑ܁D`Pf}9o S"S:Պ-E)a|:,ן"|^s v%yl, ^:CzXˢecSd]:d4ǜ,dʮ tm:y68eDqAYl(wEHfDzk:1 paZ,im8s$}H ( D<=/H#~M 1i5+7iJͧ7LfIAQP0^Uf(I>$%8!MB~~ɝ?m)9;k*RĀ1؍+[c%w>-] P]AIJ%F*H4G"e4Nruh Eد31q`SPrk帱rXKbyWnag!㓛x?%ݙ^! sZcpL)ZćZF|{Y1Qص9TP9B E{6FuePJ}$9 *9ƕߚ|c " GAm/Ni)X>J`SMUH8!IdC@\kYMx BTRqD9P1.ΠkEup&Ůoڽ'63u)FB&N b_ ӗT  X3`6Y1EW+Ӻx Gkb `0|̼+fg˳y%!18~?G)V C<Pd" aa$7#eac#=IF:d4bze91(|\< 'j_l=}ZQ&ӐLby L8-}iwF)"^ǻ&xkPZ?:܊ v-{`C/\.zZDkh7덻h[;=1X!be0))PKnDQƨq9f> "N,D5p >_ʍML!`n4"!/qbsڬI۵UIcB޸A0iTtd؛F9KNvlehƀTzʶ9,p;Fo @0Q|^-lj(ctJ@cXL3U{ ag%@2!k:dGczřtEm4!ңzF C@tIM/=y *.!CbD,8d|vsڤ<3gudsќN.aO.3 ܁`! *+%Tɇ VwG0vƓP`fd 7A֯q-BS߄EqWtCFcpTEz/%֤CFcpVŀ3aҢ6*ׅ~=xk8L hk]xm 4ߓ[d3\.#D(Ô;sNAQ$4Ʀ=;둬I$$rq87Ew3ʥ(jjHThp>-WiTa:Eo%z?#5Bbםםkt؜0*vlN9) ـ1JTTnLLUQκ\ԅ2W$QrcllU{.m)]Q (SPIr2 |-SUX޳hpFW\ޭE}z%0p`I #('d2CݯUE:d ru?=32@X,\WkݐXᛴtoW擜~Zh-֯ͳl4wھV}aM.eپ,ևfT/Շ?YX3l^͟O~v[w6/n2S>?߿ͦ)e4cXNoOMLh.f@ۃlГnhЯ#L2|zXǢzT%R|I`.ES`?{6!)~ Pzݾvn&~ulrfPm9ٖu&-3pfG^fޝ'@z4&~x6tJR> "ƈT$Γ@[}|a"s%hv}f<r's;M=6>ط ;C4G_h*#tBijt,BHhuUUkMXTPN*]u hj);WnXZ 5E=$t]nvJLr)u >#O>doEz~Ne(^<*KSX=ϳTs&YfC G=Ӎ k9ت@Banq)WMKN9kv-9R$ѲSmS}dr<,Om@SOnfM860 Q'7 1i  H=UxNnRM'<7=ttns:y枞mwl=c2r⾌z;( u*kȑGm];xp)=|4" '@<`6~~u>T ಣ42 RYB8o+@eMz*N sJڝ0=Ld?F 5!hrc>Q^`*ѭ[:x,& tNPNG+-&8afGX<`_o YL4(jSJ%s(+ld2p384ڑB#H k1O^<HfjZq|a:I\sn2)Q$͘ $q dW,gLhp0#efx 7и$LDŽ K1!B3#FK- Iϝ'L Xy `yy>G{t~,XQ|h~YdhSXZ ,Z㘌aD|q&g-òFQ$"*$ʝ6!^rJGo)x.wsAf"a,9$xMԎbY/T^; Η'B,X$_SR`,"C< yscPUq*/VuX?_.)hyEaiw3ԭ@zϭ1JJM.P8~7'm?ŐtXSkXTa`>[)E(˴ŀ ͸: i-!ďIkSC-F>%0CQ'f8r#qW1\y2g>,Pmm?ŇSm_NP|(Y|R?` XYM>{}LDcmg hp)xtk3sОl[r5V-Bz e^@ C{Qݙ2F1ec=o$dg]2tVI $z~l~=]fDRH|lb5~Sh+nL,]9T|LT='z/MÿT|f~U/_0^^J*P7/;8im斮ۚ!L]Yރ!:?sH0J1Nn!|z8sJí.:j[_ت4Sޡ vG&­ZP}L~f~vo߽{s\<'SU|a~q-l_>{2[(}(|ɕv"Cv#ӄODYv cdxJ 60fHIw+*d)IYLi a'3s{N)*1iv2T;-+-]kdo{o~z_:槣^Ƒk6;ǁz>S} ȠPZ"u: 5@SPcPFFZ {R(eN)Hp,{!|3.ѡ"w廈ẠX[;F$"3 Y--<k͑Nxp^O, i1 [hkK_(8D [AGilL';?O=cD9lbc0qc "5@RYv$qV*d:/EmBd);5ɈvRd -:#^I QxP5CrXƴ zݭz'I2ĭG : -#C$B)*es)1ͤӁ9,Qmcr]be00zk4&x24i04pMxphvgdc1a<9!'#{{s hkI׼tڭkޭUpTM}L^mzr ,4c|y;HZLƾ lwHQKa%tl7ŠB<,ƵvԭE) g]NuTZ*󏱹N,V{IIs<] NVEubRLH>x& t޻w4\{N16O5Vt4fW[PvBqJZЀ=EH'_M9J (`O 䣡˗D|y+f^ĊbEbfE:YŴ#,&|sIyٷf}:)a\64אS/_Dc~Q7NSyf} :qȌ;YA=IXIS Da6:h*6N:b+#?/ nlLɛ|L_T_hnj{tkQp +?/իY+3Ep,աXvя=XDR)N!WgxLV;kga>O ~WY'O/T<)fɫyϋP0k:P_|7y" dɛ~N.gixgmd;xFOͿx_31VKt seYeJ2Z(jH>lt8k܊\_uo\:Z3qp>܇}7v5|{&ؿ^>F Ψl̪Zṕyk `s lSgX v[g6\33<Cϓ*h}"A!H``,n10{Y ]7@mÚqַbwz3ҋӢ- Ve63ڦۋ{a_>\gӮGiDpXfͼ`O&]Ɨ7CP{_M'ccPY֓ѼݰZfln?wD9h<'GDPt^0Rǖ 1=Kds9g= -{OvbTy/@_:1 RU6Qvc/E1{tD"^' / . n?up!9t؝N0:kutr*s;Nhhd8|mgCXE C8BZLMSEZ}֙YRE%/GtI`aXkN|a q3TY'YIZikwwljWq(מ|X;ŧ0>wypbYPEkGGj~ R5_<5Q3Yx&]aTjߟNOT˙=hzKD^~t-tfY@gYȜ"$!"eRMNV2$@on]ϵ~by:Pb29ͤDPy 3P%<`8єx-3C"u@7{%xtlx[5ef7$um]oW35~LHz=֛[Y0h agTKR`<6AZ"XB#>TqyЖ]vًU'DO>S$)f " m.S0i Ә0i zZÓMu-GsՓhIE8k2%?krI9Km)fph@I ~V+V(=Gw汝b}~m*<Ǘ-"3庿pG@ڣP՚bH>ngpvY|/2osvL%,mnbowchShYS~S *1:h;w1#Py{mTj'>SάK=rޕ6$"L?(`z{^,w%.IvWկHumSʶH*3őmUQrMvL.*W昧Yo0i IWZ&,X"ބYkFb zFEo6;˜ei+WV=Xvp]}&gwgMnq6+p@wa F{"݌'ܾ --Z&y+nvZ MS͋#ie} qwЪq6[*K#kiVΝ[Q[@M´FW/Re+5j*ֈukld5}\q}&qHnްK_]T;-KM^#OFp&BK^W 9nZQeFZ )'gKOT/6&|Av g''H};yC!!`ڍmi7Tѥy5FHma2V[>ʅDygC;Gӑ y5Ro/qDj>R2'x XYl4-d3(R`K Ys6f7CB9&r66e֑aYqGkZ Hr7f,h=QX:{_ ?\(xO06~ hOZ眊8 #pƜ]W^-&_?YрmHj#"ᯠψ2ْBakO "6ޤ Kp3"اs4^JJd 1Y,-,xGI4 [r +[4~ (ᝥj%˫]רjJ>~W5I!=ױRAMͮbE3J@+~VjT1!h][f.Euۂ2!оT*ü)c(D^UUkro xK:BZ̯@WbZz+WY4c{U)TK|5Wkս5&IB".p'C(G``Xzaұ CI*? Fj*e;Q!(8v0)&8 'bn8MFLJ}YclMo3Vna3\\؄UO %cH"L'ܝvHA\|jzF9[p<5'T3Ak1gM #Q;UZA:I0!S&aJ;fa8Rʃp-1lP/5^#ZֆsVj 9Q)EC(1^Ivh(.#UBGgp0$ G sʍ؄v(\(,.=8. åh ߈UD&Cp "JFc&=j~k(_.Ks1AI 581HD-7{q@0`:Xj4vF4"$#SBJqekrr 7Þ,@ޚ隒e}㵺m#%VWn1([몭ubu%=q.`9V&؆K~-" `BK F 3<P\Fb/oD /KpgA>5wQjn׈ҌxjWuǀp|%UaS1zbq xB &D,"PC#=e/h_b\4uo֫&";Jb&"ĐrH(E#se#gGQڋ.Z oke#P$JD$hÔg*P^x/dvR8c(vPTxW Rr@eS}M+b!P]H  F0.1刉jbm)Vv}. RYB`Q' zBZ!@(c-DM\sZϣajm- )D;n|A sBzcv*gx';iN88bv5݇] Ȣ;8H;ɪ27a6#'!I~6s, eָ顴|w%h:߿~waχr֖@dcWXJSDՍbtt*tQ[V`G-6!iCRr%j L}7. Y>81 ̜3姲9?|zvp~D1!hz=].ۯz5M([Ԝ$|zڛ\EM`ʧOfF,4춼~"V x ^ |hxs[v\[ml;xkƫ_Bu= RSJ4Z]]7E)n7/@>< f0bbїDO g^N^ %z{W/44’FR_'.58l~~c(hOoƆ>vӯÏ?oA f`6sU7<ia;]wu[]Cˮb76_ όxVJ ??~~N=gYjn>L/W`@ws}0\|TBF(uˮ&/k:g`!8󎑐E"4i^͔VhRҬt,E1Wm٠P|f_A" RD2akGoX FY\ueL+i=>q'6tPKX\տH[3q6WG6!IbkK9soC.K9RC.K9t4rȈ,rȥr)\!rȥ2iRCfCD_Up췫^{g+'M9 ͣ.sp3hk_87盰%s| }4nTonWeN;ӗ'C7|3̱ǒ;O rY+냉KM-a$EV tFҙ:Y`+Z>?އsC>SuiW躲x|/.OJ#L7nΒn:7AUM@ Scȁq;~sA@sխI Liʩ-*e.;Sfef̄y:Ҭ3 +ED77L)jtSՆIР^$RۆC{x` %iK4]>(ce.wN\n9]ݦ/¾'cHJ,ie: w]=ftz1A|)>{FT|xoSIb2|CraaҚ *:4W@coFe)be&j:o"Ǒ8.;7.TZim^ChsZ6\ f%I o.ORxF\C˂wYWAž){78vWv r=\)Cr=\)rv\)C?Cr=\)Cr=\)Cr=\)Cr=zHRzHRt1B8=Nг(RԐVkL!c1AyXPsz!3R575c&e$nH12x t(G``X IB̡a#Ie*} #52(B;NxMZaL~GC%SghB,<ưUHjOЎ")DYi/loRt0㕤 `ꂀO1R%tTQyCpа8hlݪ@ۖ$őp.A*LFXEd:|"\B6IDM7{,ռ3W|5]X1AI 581HD-7{q@0`:Xj4vF46 NF~'[n@[+#]1([몭u 1V5^f 30tUE8VIO &c(aK"@ F/iD 0s ˨i(ycwݔ؂1F: lp6HF/g)Oc.*bfXes7{Sx;F9XVj ;-@Q51Qn(͈IǏJ(GOl2n8@ђS]hcBD t}Atf}K2bmO JqTQ( "2rbrX+?B"[13'3^dړhuZDE]g.j; Vi(pndW4ZCEK R0z -c1Qܴ y LqBtaE$XLj!W*3WW1A0ɵ )7\ 3f4j|sWL#,@J덦wFb2 ϋW)ww0bi[75 1{DSwwyP-"?Ò'W[1()pt bI %BЈWhCX3((dQm<vBayVqt8Wz4WIv ? S0}Wn.[ %/4pBc$NjC?M!9޿bl뀟o&X󵰚+qp+!= o۶E(p_H{{ݽa0n@T[v-i6쿿C}Ye[v$Jբn"ч>faQ!v,ؼ.sn{&؇!?rtwtXξ$alJ3}f%U7H @\$5 mk4wي歛hv;9EzQ"G̻бE\9k0@գ,ҡư<_Jp}~VSzKfAe3oWX s<$w߽n^ .P$"-P?p7[{8OO |շUT%Gwҹwk7v^cˬjL}5wM3gWMz]$l ^"iq C4G̑k\wׯkkA=^4VH#@~oMKo\=&#>߼_o4rDyT7rٍ m +t l3W`Ea7C xGs{ QwC|.jg<"H&zȡZ,6kA1GD?܁ADrD]´]/Z{41}x룙 w+1oޅJ#.T;Ĵu0hF5437yiϦWF {20x#fwvj0\%O4J|EOd9P,ۻU9IFl~ݠO;šHŚA;{rh\p?4LJķ \oIЁw~0/\`g;dѷa}GMC%cN4Ǡ B+y5ʁR6 Ǎ |( KZx)le0cŅӱQϝgOKڰ4f֨r"!HLj|vr2!F\ ܀yEdzu¦M,Iϐ}Ʉ5{I%tn羖&WI=h|U@URΎOVwsZd/վ#&I!O&w\@̓ferul멢)[jJuFT{A\MChweML@4h߇>FwGQ+p4YFQHNIe8x5 jT}ア-x'R<]WUЫϺpNNJ3^pKyA.\9ߒ9žGFbDN*WB_c`X!nA[䚄 gKmcbu` r%@YTKyQdBg={8rd>G:j\鹁hj>$.=w^Szw\2ağA/Vٷck |(l@hF3aO|n."۶^1H|Ko[~)\'(BGitczVvp¿T"sga |Ъ$TiaGJN=vRA@`Wr>zfZ\[$xF١I@ν 7f9c>|7Ǻ{>-3; ~yVV8[\/#v>v+`ثENR FE`6Cd/2C gU %mw@_$:nwdz5gO/V\<֖y0>BbCnlKKm]2ة1)[] 3lcV>.'lo4SgE:ٓ 5 ,5Rt*/ٺ>r^v#=e {⣨"F7\n6O!;(}3^{"DYr 8e cȹ29 x Y4R̔ sV@TDBAp%{F^tPK@>>Z KI9X!N@ ,qBc1?X"VL%<"q 8<o/,Z3[a<&E+Lt>MAd٬ﮍWc ~@lBَk32hgŝL=m ۢɹt;aQya7ocqs`mFkvVp$S>K aۛ =m5N7~'Af Z#.yn xcA[ dW٘rU7|VL9[V3Vt|m?2m@~E >呣ZGWrֹ8 7)#J$ = ^7ˇ;}7х0zx`,+'U%g3wgγivQn[ɳP*GetryIt/K_m`˽L-ͦ+s[fe+qբtWdjuN_AdQW) @dY崬3Mm%h]`kVhANclB脒 b'o? , J _cra;7fgϟM{eU B<5)w: /$)Ҿ+ ;0{ت-CD>v (ڐ1l% u`YiPWsХ`.;u09;괍Wq6[qg$SJcIB` !da-7}ye2D5W:tt&=_ 6(ʩSPHs}E_ ui{._ma-zm<}z83q,<7Kf֮.cմVs!x꿲bZ.DE-28U%^ Uu-ѭҋm:*إK/D!2 Nq''w28{ۗ+dԶ lGV)ń;=|ڞ1n{t 2C[iPĉՎa[]qcEmn0Ro7qutvmd׾d7+6vdah7v0Veǭ6GS<;Bmv:婎ήjڄv\9BT`D'{E-ǭ }1:/2F  ct0q,0ci-DcRP >`$ (Cuhplfk`wqeVȍW;|u@^ CVnsXf];̚5?u7#g~Z~O2X  L 뮊Vy;;ЍX;vΥpy<@l6}2irx$&I2-RhK0i/Sݛ?0^ǚFJhDpgaqqc\xT䁬b3_z3@ӭFbÐQX0Pcm$hHPFɈ5 R&HRFWK/r~*ߝx}ء ;ASkG#=_J y,EhEg$2ɍR$qቑ8 TH%A[umwe/=mwNjP%1}@qX3|d7o<+<K_7P /c% LI/r6 VU3=  - - "Pg KBE1s8фթXj^jtNOXv8S~$t v=H{H'adMFrߎQ.[V||IF }B{Ō~LI&dDGe| ^M|}d.XBC5L4ֶ0"J$2A1b0%dX{W1i^Jq}SAfjNUuR6E[v67Vv<ۖod'҆ <9LR{rMUcExF_(BsoPE79C` & ؇ :!ҋhޛcÎ/ qRbnv6uidz]쪒+ve,l]T3VhM.;ʛA寲kc{ú/CXj^NLg6žQpi^ d:li^8MT\ (4t(7<1cqFUV Ōf:=yWXǰSIAryXi[ U^2R@B"0LAUXMeR5BL!dY|Hed!XȣѠgcFFƒ }vAj |NVE[G:Lq"Kڎηs拲|Y-}ƥ:[okzK}ӣxmeyW.hA~ y2]jtv|Эn&0>or՛o/\Y?xzE5jۙw鬞L_Y c{eYmn7q=)ޖnݾff|UfyɂXh6_V}D2fmNMU-nur_|Vm62>FbNں:Tjc;J?Mgד<=}}yo_>9?~sc<=~}8` {tV}C4i 7zGkTO{!>R/i֥j/Oft(mm}]ZMͧkzV&e_,6uU^sn۫,Ui{=D:S@ebUTU!!HW mN5)CAmC#__WnsgL9 +G0eZ j b'C]ĤQӜL}Dþ^6XOhsb}}{OMlm[#=L6lZek;?Oo:Kj͔-+Ř٢/lݕ_=Z&c<^J=Z&c^Fܼd}b%_LV'.ֵQ+#t%o*OZwyk~zh `X$ڶKu`_•f^ʹ)v5)s]vYL)V N+lϋ16*TZܛǷ58Ch}kkBo!I* B/H±\ ڄ8~f()i`f->Z)S'/T$dXDQE =[ g>ytq 50ӲE[iK%$kSުA'ʸY],BYP4ˬnc(1)/ "y"Ȩ& A z:URUCj@R1zVL[cFe-XA&'tnc(zk{ڕV=V(=Dw7x͂iK@]a*ҵ#(%C_&HȘTt2?BNHRZ6I))acjgZ?B^U9-YWs*yAYTRo=~71^BdYi+dqh?B`oL2Z$*%k9T<K@[Mc_Vu1́4x(/uC6VR ENXV)Ŭ>MNl+oU.^>W, 5I %%PFǂZkZ Pb~Ah%!t`>Cy8tdSVXcnN}+? {++ASc O_OO[)gĔ*yiȻHjVv&b c_Ѫ`̓d/ok[w`bC/Ty-4)e5 SsT!:%j BP =99tJEXoj ZVX6 CxtH.#Y5Q1A9_:YrP9{/e+MRP4 y=Q'/jJ}b,CIc(MtTfAhfZ2rCUFfi =Ξs[DƯ,bf`o!bqb.m53mVY =ozeeKPT\ʺJd~ Tw+yr*$w\PgUmP^)ݛ"!Phy!B&q A x{6UDa颐{+eK]AIc(Mb% x5 'd5Cx:Ӫ^m_~oф~nAȾ=Χ3>gXb,bW{ͯeotNI' __-ͳxU.Rm &mEU*Z`'YA*ʒDLp{ !%RXXHWht>F<{ayQdW%[= )Թv  _V*)T6^٪NT@:H%kH5zBbZ YXS% mjj0Z8WStJ do@*PUߔkLo䥞{siԯlOZ}ff|qz~mrb'U?-(ʶbc(b/AϲXڨmuϵ|qL?\}z6]۽-C]\MԞft7ʔ>}QbF8uyE"IbE&q(Hd%>fZ_)O6XݞwkfTՉ I$*`IS6Y))9Tx!͖,2&7~ѶG6W.fJe\I1/;6s!G#* fŸ <'-_J.H#[J.*Kq$@_*!eT<+o4%4LcАT3Z*5Cλ@p 7+]Zr߱&4kIieT/\9JYNJ{A$8kRH`.e)}Q[dƘPp_\+MEvMn=!sJ _="?c}7_e7IΗnP;ȴ3Sѹ "LmT8"ӱ!")HB4$Ph9j,ҧh=eogvjl@\Ҡ%C#D8;Q_gZPo%ąU[A{+X@"YhPry˘.=P/[Kjv=sfEdQ<_rډV=}Ci{j 8CY8FyiCehf+&qxY @`}&Q͸\ſ; Fx%ȨhH%bL, D ӔlrhvBF+2[JО+MynQ⵶Tx&RJ'A-Yki(g\lձQ[)S"%A3.fǘO L: lDUx|/_+z5vGgZzQ㗌&BXǢL^8JNsQA֠V;;^n_JK@`%OF!bIJ OG+c#΋mk t[A;Q#B0nrLv2aJ^`yˆzm@Dۥk'myDpm<|ѓ"|K϶,|~MPARq0r@dԀLYhicZtўx, ׷"qȋcAұs>L\ OCH5Ci[%d)5+!l2y|{Ȩp{7 Fz(ˤ%m!2EF:FLJ>[>:.}X[M-]B.bҚ I*F e:g&,eJ k9vu+Ar oVϤ9*x/w}~+;YoAMժ N jdPYxMwu]A?]<;j']5t F{o1m۝VJU͔*X:%)^;[no:Ą [-u tX>AXVtI$uT:8Օz*?o0ԫm|SRR2sC7̦=0cpAb<L\ gKМ窍P+dbPkUmzGħ~'WUNd*Ϻp>:{R{04`=0u}7z8Vs Ç󈿦ܻ0|JUL"O]fd&׸ ry=8je4?=s090bLԬ&đG[OOf^>b95o ܎m7[Sbg~& \k-q/)L-78zW.R]TW/ 1M&'Ǵ\% AԵ[u̅3zųeoLgvnHޙ1#A).lׯ< m:؆-V_ȩ ^P:( 6h4S`XC4hӆ0;=НH²To$x2@(D(L1Ѡx&()2|+:PnFS# VE3?ר2iF! |rcQ&/pEaF"QA֠V;;{,} 颎JM,yJ 2%BDT%1+%"F>IsS(I>T;/ OmQOg*H}#&Ǵ`'Ƭ 9/bpbVq2(6[2$gv=XqOފ"/>۲ZXєIPʙ풒Q2gA) 3:ģ=X]u"qȋ4mS:}cNЇI9XΏ!鰈X'̮c7Ϙ:4vėU@W F/*2޻Q3aVP\2j"LIHLRV4:H[$ôWGV2Cژ6tcfx椈5pI?UKߙfBhY$V益u޸\͕rNK|"*̻@ɀRdPH3z?C?3{-J%@f{OIJ']RAn]p݈ljuDzYZB ʤ9%6 ŠM~%e%e<; kͷMv#RQ)wq]k|5uwN_VPnnk.-/G%;jZOG=..OOȌwAbp432ǻqyNX$nJieT%2%G|$,%٤ʉռfÿ:Z -3u8{tMz7<[hTqb[u2^l x0ț0d rN-%:AW5s?"f7[MM&_ByNc᧿Bϡ{q8a+whaqػaR7T2lQq%ceL?t-6͔ߔL/]8Q c6Z`}KzI-IAts_hM7J׼Kzx՜ho$n>u{ܻ0|JUL.r2'3eௐ˝&8MfOLLs8o/5 q$#O^4D?M̼2r??℻?sokM<L\7FYPte{"9|h>5kj8^L PnrnX.|DXUk'2sA 7:Ͽ6ErN DZOX]{oH*D{ ~? 9O2060i %$qWMR$-Q@blUu=rRYO; -h+9)+0[iׇOw?e9VO rH# be}03JM-a$EV 1ִǥ~=`S7])ͫ"qYh0_^܏'maʰ9y-):`(3*h2:ڀmT 5ѡVc^9'7CZޕp!y&Z"УR6+Px=qVZ\=X*qؠ4߹b Ś}.z1uRKB6ίEfHH ]9m,,vdB۠Kah_K8&߼ʌcgmIUZr;Afx^ǒmJT)c̱|29&}uNNԻ+#ǚWjҜ-=j; [^F`%7rQucs[U-w4Q۞h=8j:'9q^R@J[6gE(?> '^̀XGu&\tOB0mq]9ݦ l]r ^pI{ ki)$4>Dx/Re[Iyv9K2WX-I4z<%Jٓ)=NސYVN zQ$SҤlҹl \a)"\Ү2-=#vV4i|2efcx%\9Us;P;B ʱ[խY+˜RIl_yV3xn29h)$Ť;ߥNq&els͵uՅqvHK*ߞ" `c3YiV 7 .!c٬BZ3ɺS-5æ^X5(PSkz02$D7ޜ}x6ױ5xJBFء# .E%`SPmL|*[7g xwZS̑OܟOH*ߍFA +$CsKFD_Asߡq,B3IRMSDLbxw0X?Aya\`=,YyMc% Fca#\2NUYYh6Zf{(7:˧'] nP演vѤK1k^Ǽ!%o#"*Ċv4J3+QH :6;my+8F4j6T3dCƻn3x)m.m+mFaSTJ"Qgjﶟx`[j0\v5r1R&gV'܀0GU)nE%ěJ)5 }GyxyL=TY=aM|6fߔA;"wɫ׫vcOg#1h7iIoDڅ^|}sQy|._0s{Y7%T},3|紾WzwcaAez-6QqVu2/_/ q c1fqh _:xax4 u<S I7^m$J7 Y"mܳ8-|-I6UU%n;]ntUK_nbe0~ x-۬Y1P՞P# %[o~:4l*˾f3™i@U= F,Pj3yo= f[=19׌K2Qe$aMGSq\uX<ٴ.Tim}mJI΀PԸ;qd8i6 &>ư4a1HKX!/A& %N;-sDD+D)"q&7;n{Td`^6DoR5D;jy,-¦߫:ΒB^w$8LlǹúQF]/aN(Esfͭ$G!tpK4Wg;C/6EQܺۘk𹦜cB)7j<۔CGEdQ0A[R.k%#RHD:Դh0 ̻u^'/W b!]GuE$a@;'YmXbz՗~P.4tڞ>9ܘlNg+-{HE{>E~D'4o h(+w_sq -~ap-@R)6VJ]$|UW)w$y:y-nA$K3@D2ፍF }W@0!R.R\:$UX=\$ ?E"QQL1g!ЅL6eKxnLT+|RB'O)-[V eebj-ȊTpt as0uV6OfQX~+|,n\Uي` {WRs" / fBu= R{w#um7 чأ`*&I> =~wݛb$z+uN"”GRÑKd?'8GO1x(hCzvz7W@o.>CNj>]`. u+0 . 띇NPd<i޵mu MmYW ޠ_ۜrM+G>nąBs3Z0_nzctxiU:${Jg~6u-*W=_GńQk_l! 0v@.!6w'\1HhyyŊSH"Бd1#HƘ#mC VhC鞗Gop"h,)+F,3HiI 'D!vZ99*ghb+gݝg{`xr E'Knyt[ϫI':ϕFZܠ(~կʣW*Yv>7b3$^·3!;FIf/8T95,&P}Z+)K /P]z/Okͥ)*ڜ1orCȽg-Z;X /&( ^!whb#NPaieG,¤cUP0 2EAHMLc1>"'^q<-ֆNʛa:4PNV!y<ToO>)7"ES9:Ű=8E}眢!$Tvh(.#UBGgp0$ G 9GpB]6) p dA8{r si4p&Wm2|2^Ĺ`1X>3ay^2 "uFҏa`q%ܕxnnSeiF9%BNͥ?iF ֏)YV+Uw{Lcʽuz7)nw&d3 ;L9y[#VmYӓ#l7]/,V?v?~w &uD9l5y^fެdݛcK)PYGH(򹖜倐AfےHʽqb޻4X.B(q$v:xLu C!f=;68R$#Z{1Ha1X@Uc^T.SdsFPs p6GHvjaV({@BA)'HKaAB|ه )ogSlSdmX4c.[zht jdF8%E]I5j;W!KgYc٘knp+~67ބ;^VF+(|I$uTȣ"ܤ(jwlI$|gӶpLQrDj/EV8QnU3t3O7K˻v8Moks<xgp>5)WZ:oEJXp,c9̞=MR[?: {?ϯ?wtWW6eoR͟/C8zٝYx">yn>Qh@%`Wpd\K9fQsڙD{fEw|Wq ʗtG[C K$. ̡D"Qş.B ZDUUŲ,GܳGeH2Ya$]<(TFr,Ԯ`fQ=!(LK?}"!ؔ 2ř mT!@(+h3 ΀O W-JY0Q+L@ 23Cd|QpY-APۅYQ@ bgm]fǤ`s 4GetI9}\3t ZdO"#DC@eM𨃳3c1kBr6 *RTjJ Pmn gRTjJ P*5@bvbrέWCJ>v< >5τ"Z PHDj+Jּc~GudrLK)ǗOvr~7?vkۜ S;֎cn[;֎cn-3T;cn[ѿv̭sk1v̭sk1v̭sk1v̭sk1v̭sk1v̭skKśIXJH36L}kwݷv}kwz]o[;o[VZ}kwe熸,r\\2VNhP⤦~.Rh7j-;7>(|cVz>/K PT0&`^S@?U TbZ^fn<>IHT7 {G7/ƢbXM\C=0mފ`&4l-^A^/6>>J(6b#5UYhIAe^jD $Kj5ezcZZVE.\;wJŝ.#;ąUTauPFkBT^R+R˹ ,^ kg\WnzDܤ2取ɤ_FT'9. <DРl2$l$]f)*0I[Wr j5R]Qp3H R>DrmUTrk&/TJ{*)N+K"rbGk3)^kK gilR*U |Y1uOf` G.%YEA 3ç9Et>g(i^IS{-?e;eiN!mo Z-kTV _FGWbIS ۪iZsɤi{ujat%(Y0ZթJBۿ묊>qSORjT7eCy:ƣ^5KchpOEVW{s6;loVL#<̪@ƩK85:AsB;(Yx +)keofO8h1W_nSS}:lYANgV+Dp)UBSHyǏSpOWo^].('x2|rAn1\S6O\wn=g1b3G7o( ZaKfa[ 3MAyEg+?%w?t|'*C 9A(Fߒ!jfH6yA#39E#d>,R(!^G3]Kzu@]IŜե쁙mN~l+TP+J!/WB fUtst+ D^ !wS>9(ӭ!-=YvJ.XtylJ+(f%̗#u5tu„UDp UxUzu!v(ݶ'ԮK}(/#i+4=x |0S1q.W7!:>@Ή+-מw:CnJbewqF<` `|\"#czL,aw%ۺ,]%0fdwn%qorl~ ^;cǻz_q4md(=wA'κPZFN&z(ޕQzy{؞ (_ :O>_p77l6a(ʔՈPpLYVk>GST{8tg#5_i煉M~xs)w}< ) .`[6=\7N) IǭSh!_Whj{(J]l6]O~6 mG_yf /,m^6~?m͌y6@_ʖeZ>S zjf$͔" vӨ U*V2X`t#-rͺʒM,QCBdjZs-J6&[mL KHOee9wCІ(lȣ_wX,lzԔr~C=`K.~WrxK_>S˛/4BHS(>kh5{C͞w65[UkV'ЎE-Jg3^\c_C(Zzw@7bT-UZ9v Z^5tq:!8r㸾h c6.nROۛ3q9m\+ˇL8h b>!6L~G%p8|J l!/w_kTcsIg3V _'{/6"Vj苉cW:^}[q8pLPk5'O<O8/5j|4JZ1&3AuDH%=FySHczVރF)( bLylN"klZbN5lʕ6I"sEdvq6 }dr_^9>bWL3q qNX _# ,)9b2Wc4 izqc=yDj8cm_[8U˽ #G*Crvq's׀BHˆ+ez8{!k;niCKm8y]lXg 6 4 sv 6͟r GY鯝 2Z f =sYﶙū]ϋ}/xa*^O Š4wS^n^xNQUGzuo^Gڿ9~su7g _p8).vG-T/wo?QqDi\a]޻L(Ko_O:t>@W۳W0ϸS*C`?FvlvUg[qnR_/Nx󽐾Oo{Gv hw7G+\rnbqCU}΋n".";Xy_0, q4 a^ӎrܜ|\eZ#=X%7zRqC/BdΔF{HBAI㭔ϩt_?7xe;{^}~ zrN.ur9~S?/e U^"hJl.xZC>[T}N>kOlԜDsFq(Tٱw=D0wuIOraMkck<&JZ,Rm;^Iz$bâZЙSy5GzK{XbDɻ$[kҌ9#&Z.nPjfLc6{ ckcbq$qZDL5#Lã ̙nՅ[.k J6|YqlVxbB1&ĐRцFTS{L5)1x!q| wAd~`EN߾m߲I1Up]E$\Y2!YCsE>QB\4FT)EJ`'9PJTr2=$5)Gpר#̇w2|GHcI-{$qu~/kkN0!],YA-ٯi EPR"Dq`exHV.08| Y#PjT^4DܪuQKljD1dk D֥@]5+DC,/2AꢱȒnQB,K#PkCuFHFiLHD,x&*ծ&crMc9g1`-ipl\[F[U piW}(ė%$܋ÐG \Սap-d]5 ąeDC(mR&n!/D`E,QـjXk/$R ФQ []9,]꒱.bNqMcrx8.IθwTbd,iذ29!&4 BH$S^X$+䣔 KM(͐PwHoZ2IU@YA1Y4\Ln]DZ@B]yBBQՀRRQ?<#7 cºXo}B 15u) ex&P/4wйf(ZHx"fGªaB옔?%cG< s% ec=. ׭`N@X +4poIAS1Dt0GQF[4yՑe3 0|[mHPS. :eWhb΍^"*e^Fc8Pήi%B7ӱ~noaːUNJYc(ܢND B  fs?6b!v{iៗ\ U]-B&1MOY oC0RP8r0}J_` XD@:P{u Ç<)Ԅl _~ZPE(H&jZeT^2Ї8˭aFrzOyILPD _ȚHK@N* n7ckcK VWES Uߚ~ɶ _EyFm ^֔$S `|;ܮ7(6#>b o, p`D(cﱨw@d|"Ū<$D \+:&95f*H-=U/-hbn:L'^BIZvRbF/(3Vz+u8-^؃d1 `$(YKⲱ6jdԌ4 < V;M#mk-Ƃyش* KP7b@ ܁ nPB@ƶtBsuh-^'AnKctՑpiDFvTAmJyhGɏGo4佒"ZfF[6@j\&؀$>撣Q%03 ڪ F{ e؈yp!{`7+U:y\0_NZձ4&-IAq@Ea桂hI5QA},_ກBDF(9`%ՀlܷzF^rrr~M}3Ζ6q0`ctSk_Q,wilJU7d#t~/+C@ϪERI S|>J#'VƩ@+O%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%T}J`RI Q䟍~n J@S -*N%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%TM%T}J =3`p|Fgܜ+S M(>]ko#7+״wb`gAXn9jٞ"[RKV˒l.IM"""Ζqǝj{.~0Lp zͮ8gwgm&p dVg+^bn^ ))))))))))))))))))))))).IFpB ibT= ޗL;\EOB'h'8ɨaQo(*3 cª]qUdl2t&?i fE*RY Qa%˽ŊwUbITj*R*ݻK^ξ듽~|Cr?_AVw_TAtʉeUOiKt.m?U W:^_{rǭ/uipQ| 0g`nH`++ڒc[8XJoZy.}_N |H Q$JD (%@H Q$JD (%@H Q$JD (%@H Q$JD (%@H Q$JD (%@H Q$JD ~(+6K"xZM@o+-;@ E`m`)O֦>X[<6GkSR`m0 +m`gL~}7+`a?vp\A+-|CW8&7|,zvPEC' x-4ew8_m ڳ#aKoh~tggnhoȆy~h >OZ8`dAi17.VټPS< %Õxz3]!iז5mQ>^,oR"\[iǃrYdz \RIFQ;1.PH9\NXH@Lnl.X´] iX"HbôAGVOׂm\,ϵBÞ$K:t "]ulšeҴiIi'a{PiljUđ,ĚV\r$J-^5x2S-<5ͳ{.ƌ^iJzsLс#aZnH/af_X5]@{_rzh`}aauI% LtH~me9ppL遑^k.#Ez=\V v\H/ LH=͇u}}UFԍd+X:El+8 (i%u!"X:4_77:%~Vb dٵ rh⧛m3,HXa:0Q\J3.s1Kl&iD@'s*a8 Hܻƾ%۫xҍA}pC^73ڐVf'޾xf\۴2͌c;UZodỳcP:Uc.zgQntZf֠oKJfPrr}Y͕7G[*~>DMޮcGͼboSϦ ;oEVuѹ45'ݟ]_k k~ɫn}O,ϒ~(\ 614z.~Ϙxv__}-NWu¯֣$GMf&a,XUvn3:Ó}!  ~1֢ L(#G愜Pu&g7=:HoRn=:_'Y?VSڿ=8m%Izsz#d=)Ȣo.VQ~([ bukT Jl&\ݢȽ|u^^Vw%3)sNB.XI)^Y)fCI'iŔvbۉ[F=kKl AoȎv.ָMUhe]Gg &ce tK=֘,O2FTYf$:0S. l_)v]%_*i닲vGv#Ѻ"곈'Ra5>g@Qg)Di Le&4q4?_"lW-ø/zc(wGX^o j}^;\b9ua=aǖ-di8ؗq6P-S 6:^&@&unRA-R6x`ɂ՟s2]Ș!GtJW^1Ǘ+ZC{6-6R29{#=F :.Agr HpB=|Bgs+gk,)3ap:'w`K`ZFS#11)Ǚ46͕'e;Ol%FP=\2lH= ['8x(8ѝ9EMU- 8cJlN+fL N '&MV_# ۯׇ={6+J_f땁x,`Q1EjHjleWݼ$ْ( 2٭WUj[}[i8]vz2*JoJ,wcċPbT񪸰#OO*N= D9}[S.ߴ4ø}6_"$N>T6cEzv3]/ՃO/^+gA*K`( gfk"DћYhxɎ;!M3)8g:o5OM4 1ѸQy$WfW=qVFϺXdӬMkEY9f4'>O|r+#~JOh11qeP8l0~p w}y}?~}Ǘ~ >oyR܁YT i^i?1]M- vOb^N\rüGqeRVח`s돫EAԬ`1(a+6y4~>D_nB%RKyX"~qܴ1Hѣ HI5n)1`ĔT@BȀ.yqJqX" Yj|*I`. VkGA'P^Yp ~;vPg8W-۫|ΑmJCҍE zU 9L0,|ӆ;j2ΙܬT])D%~e@DС.Q&#jBGT-:~rPBDR%Yic`c3pv+} b<>W8\ gp1G3! E) Xym9W!7LiX`פ7|V9&JRL3@sCQ᧔8( ADYIWGJ.mH` $Q>+5's`HqZ:Y &h t'9POK v[Υ w\0dèӅ(Lq IƎ&wonvrR*9D2k=2PW(C p2F~@"I";^rB^S(.AT**WNq*$ x&&K4T<2HFuNsHN:ˍTfDhl#FAUi@k!hA_=9h\g1iAIUYk_kWW3=޵snr!ebc+5#T-\%*BV~)obrd # ELK 2gB4A HZ]OQnZ$(SBą6sTQ(=FW( P&&tJg$5LB#S$AgI ENYgത#9ޟy[KD혫9M(P!nCB{Bu)Iמ0M@&YOO猺~`gy;ZFfZ:>Rǂ{ DU;Jtjujٰz;cfC RO  9YF )d%!RLF.Jg0* 'HO @< `+# cj,:Mt:mJ"l^!}r;g֝GaEqG== E%>1-k^[h\8-d ʥӒry0x=h68?v)o;6GeUlUoI_bWۦovXsg8ִ.;9wȖdS)yD)m൶Ǟ҆JKJ[cOiP>{LiSh F|sTA$(gQJOe-е<×J\J- CLmd8meۏG/^BRAtPmI@'Y_ SV{1EO|.E!)RQm-K# ) k,whϋgfq0 ͮQ]jM%[k&Cy j)tXxЁ vWE?9-om'|V=+Mg5Sq(Tb <^|Uʳ;d^y90JDWo\>)6΢i2"Ǯg4)]a"xUqӟ81L3Z0hh&W>?)^8yB_Cߥ@+@v/?/^ܭ8YٚtfΫJ ~Q=ٮEWzNCV;r p8!N }L8Gtz1q"% Z嫜0aV-~^{hN׳XAK^_˼u+8S`Rjmv5w6_lbE`݂o&`NmfHhV_M^y >4Z)~eK* _WRA9|w6nJd}8n|?LY<`+EJĸ cAHVԧiEGVĬR;SfBFgWDkq^wO!A!dZ$b\6hDQ -H! X83pZMJzRft6^;eW]:poڡ>^gtUнEGPG]df 9I a^TZSBVF8*gReI.R3HR^qD<#hgzY:e.=ZZ%l),fOzy֮[jBgF0(5\< -z0D=̺7(wF~lzU}\Cbl!Le溜vU7斵v+!8V9UÏnOʂ+i@ U.)xʟp{WƑ /&ȺY;b2cv:E@@ԯF@4lFH$]YGVZ&"HG<.BS 8#%hдf4gc6cDFtT"S$H߂B#ok T:RDq"g%|QegQݩuc Volx6vRBr䰫:$N 瀟 8 tJ`U Jn)` RDy][r9}ԕaޭݳ;\&;(Yqjv;;,r9}tʍ+wluyVݼbܼRs*%NU#!^0I<Uj }jW4>UG@Ml|J\8 Oz̩&t mje ]jFJtZ+vWg\+ LUl `eeGt5[Y S٦!&`1oʫŗ?(+4hy8)+>y>M]pQoʠ>L7ȧ ;p|]uzs8}n=n ]-ڙKIO#e-ߴ|s|s03& AS9ezټQ˯[55ݢj8/r le1erh _]Z82S<:ZxS͒Vz5!v(DԳDjjsm& "Mqo.wNܒ Vm.q$-i Bq#::T􉏲Wc>ĽJ=Xᝋg&z$USڼ7^τ혷zcJM`fسk^;ͪJs:фvj_j_iglg´Qq868i7 kizkFVWT$#6(1o< Bgl:C-j;#ϻCzn# ZUd;SmvmJny ic1Ue:IdZRx ui\"}85ϞHLHl-40CYstGwyo՘;EݬNxε'-qdMQ풀pG8m4F5JQ#epg5蜋(Cd(J\P`Hnq:?,SjΑ)E+&nz{sa&JksM՜SP3-}(|g-cR@ړ$Xtd(u!HZ=^78N@s`sk/wYYGk)%xH'UR4T!GD%/gAs\PYi y}NemdvC]E ^J@TR`0C;g vd]M00SOv>T1(K.8\ 8.}0#d6Pҹ qNSh!*TRTKҬlbBlY/vny \I̔u>v}&沯3{Dr2SFI:*ϝs~(373{SN1K$'IeP|,k7`R# #WreGxzW1nWg$ZqzBPEbR(W+3rH,O*qYƆEN?~-ןCΖI5J8?L?U+w7oե3 kA)l~ U Ԭk=ʊ!e9*UP: }LQX9syn^9[A r9skSf2<휣Zqywc@H]KնbuKuͰfmf~GF ?(JX~a:У6imN;Yj]_ciS \> }~~kzE;I(>/ey5evwnp9ӻ)#e㻿(u#0#AI^hOwnߴ1M5M67ͩhՀo.Mv7Ҡl-_|?vG>}x&91*{u_::a$v_u|/SrRPNi!vSeݠm6Ң6X>=[<a(&l5qN(IC i 2H»y.筡tlt+Jv^3?Nei2Q$6R lJ9%~KzN#:FؚbscT=[ީҪP?#NN礫zUyu+(g삎-e-L0rl)S[Tj[zǖ pCaR.E %aL>[^HSC~EԁiZOIrI@CE[OE˫H>$"@bN0T1h|^h՞yɥKblPBS$2Pb6' xbNzlSZe2 bԚq8  ZmFm6t7uCt>L.95g""2H89+)ť-6ZOA,xmW4T9'NѯSL3K-gGRZi^\?BpaC{{Wh˞u-XT' 1N:%b {6M"Pe I|զ'9/QOK rFv.EE6:]e[TPM[;#IȾNkG8~᱊,hF/NМO`O'%t>q׏UMŰBHz `>+NBa"VG+ܞ x1zaNo'2,%@DжӚROmZ 1AxU!2{6tO]B:jUBIWC,s%:o ~wYMdn~k/2R y_DD&>\SҸ hoIrnf ص27! e*4a1|KZ@нy]I:ZO:Oy'*&h5gݏ IiYtFlzlə+'P. WcЏ;Z3љtFãT|mT.*OkwrG^K~4*_MSexPdϹ_; $Rpܭ# ה)x[|EYxbbkŻM9Cp'S=kތ2pBzbDc!|2H[SQUm֌=X@lQfiCz!V\yZSeS[pz}"@Cؖ0|MYCҳe7˿"lcMt ,&/z _x= E& ^iӠ5cj(mq%|DN^w`p/{3AUK{X Z\|iLKMbWXAr2*ȔbhsDzqj^*ONw~"[XDl$3E U)bXVPΒ2:E[z]'N[(IY/ ]20{j.֧.T"\CǦD8ΜKXG1Flzx5G*N^yE4< iNzG|,8\׫.1y/i |N?uqi-H{6kA酭Fc,VʧT@]bňNQķDfQ$.،F9 g[L8O7ݤZ[XG&bv)ԃzNڵ[:5kA §TW]qB GgVF_FW맧6L:1tׯzBYjOn{|?xMzFLJ`fyvun8t5n뜻{Yx5f9}mkR-vr?%cK^ x, w($[{.rs@#P\$|»O8ܷM,SʁuUD Тmqk٤.EXqKZos(Zs(ky,DdL+![ 6j ^FAba@eHYr193Jb 12:mvo -mw޽4P ~͑Smv=TCx̜4E^ࡴ,X0p.=71DhT ;[: s& p >q1qcPAKGDsMh Tzt\FWk9cSr]Z,aƜP9L]WS+J}wʹ6@=lEUj_r0i}>W_;q;CbRi2@~dt}4=m,׷˪Z@'Vׇv8 Xu/_)2*4\B%O:gWN7BHQgh@ρ~ Q SZAȽJWDaYiiH&aʁ !b-IcVɓq {kiWDc9c)I6ke4 r!$}@H:indH`3.ZJk)mvh0$'^̽ ;PlJd뛂=ڗ4WMDZD+(+pWL2:0إyFkd#P`נ IWYd 瘙^Y`Gvo?E. W>F[<~_u1 ߲?PBU}H!ݟड़/Z}~D"MWt)dpL+Y[%tnN:l#K+!C#29E!"d>,1(FGb?AJP<&uM,X:d0hʀB+'\fF c܅٬O^I/5l_ϖJ3P{૗VdF7rĩ +p"`Afm yclht"z4݊XSnhoiU+bal2jhl,o[PJPoɤp-?_{o6;L(Þ?}-Fn9Kacw@Z4 y㎤m:8ڦH8S>\ފ>`N:l>_oˏG֗ ԭKN;^PK귐h#nZ9Jٛ3 tS|K B9_ݣ/#->eդE#67JY QC1Ye\-]]wKK(%̖0G uK+ ŽcGP0@mWbM@=I66wf!i+4=x 1rcr .=5tuɱ@-L6G7uO:Em @{iRbbNEl2#Ɍ%d4FIO"{n);ͪJsºфvj_վ{1P6ZʉiQaېw@]ELcHnY4Z0yEGA,E 57T0(o 7[\5\|TզUG]jR(MCۜ4YM9Ij%wwu  9g섒 œLRVWSI*{RVR6);L0J741>?Vc 5IGɝ׼Bl**dGɴS({̉?τo?n.ST~;bcޏx{ؙ֊g<^Y:΀!*d&2#*|VeyNJ r$$ F]L{Z2d8 LG;tTI$Ȱ&GcyEh8[bh*sJ@*#BZq9%P*xš<,QuKw66剢Q j̈́N89i:F 8R60/9*" 4G* ->t QP%g10FN2lTh87D!$-UzNǣX#TS?-/n{^#HJ IlE0L* %N]ܢAt-hچ3LȐfчIH19 )Ǣp>zF--`[% 9J.sUT*;pZCymP.&RJNj)sJJB`}K)?kg"@\9N)tYRND<9s.|* 1]H{S|0ο~xrީ]]/xʈeY0&yè8#0 θd욝]-%̍FF7RV{ߨYr@HʥY *4jn 醏9XcԿ)s?vEN?YJwkW3.84Doޜ\]_@%O糰(Zr:GB1_?0}%YqY ?ޫR5yLѝrCkw?w~o\x7-πr_^\Vs-gKU+ir%PWZX+GD15,X$߻n~Uf!\UNrժKݯ:*iJ#eaE0oGĿr@3O{1THfRM8ɺN^_:O\~_׏o~ }|[:I*x/~~ikZ֛.-6/MSO6|u%oW|=# n֚~; %naRC{9>sNc'NB`_φNLQCE@k=*\Ś SV.WG,km5V=~ 3l⥆`B$yB[5GU9)ϫjv^#euGEgzڃFv*1xV cXg=Yr3 t+Cnjmʊх.1K6f 1O ӽLxzCQM{["+WU1U @Hhjo;&abґ]0Z;h[3/rgP<;P hظ;pRb]ia4-`69ǞR4n׽܄Sd?,@XY"q'JVSm%nckݶ٥'g5џm9ăB=v]Ioc9+\Y%%Ч>̽2`tY)$V~,b$ )QT=F0/g.( I"z wI8SSMYtbMA̵PIiKz(o4蘯8%i\=Oy|jtr>gpc!Yi.$"mkQߎ:< gVكa|]/(*ՉT*U.ԦlŀTrPJjCT~h#5-4$73I?06 ̇TS2Sڥ[]Z9( "g%0Ĵ1ʚ11-$+Ѩ!D3`FKe6EUZ Q/Z-(Q(y(s.hGZc*Ap;@I賉 VPR][]LX=r76خ݈g6S[ħ *:$Lی960iGEӘ3 Ød9d#X!x*E~?^Oo%TRE\J1A0rg9Ut͞⓵ijXļGx+F΄sO[e0]zREnWDՓttmDڨҫTr3\-/%>S5gT(hnW\,71E@LHJe١Kie/E H%A?1./lEn3̥ɡtܔ1R7q@В/*ŊAcG\k.K SlؘɀS8EI[h ::6[:ܮ jyD=ɹ0_-Zo  c -R!,&"jsJ=$Σ @ƌM:{c2׹pG*X;>;}0'I=Xc|›6CbT rw8Mxd4 (i[lÜFA{vnu]28 LjH-1D]xxP7"C5]%,V{1mq$ }KsaYURLj:#j{hQ~^|&}栚:吾o5u:[~.(U۶P UӨzV9{^q+y=xy+;W{^q+y=w;W{^q綞W{^q+y=d]&*/sb07\xSLӃ07+ž;fR17N^ zD/e A2^-SǕBEF)BUߏ44#gwej&ϩ}elbx?u7+oS?-CKU3FS{~{LlHi5_Y%vg$ u_o/G>IGJ3E KM(h:5Υ$ѓc$T<@J`s2¨ PHH-k)'vkek, ]Gkߪs _8=gX F?7w [STf!IU)c4'aEM'Ad-mu9tV!] !RUL)hd. 鍀#YPنЂÞ/;u{G_Zg$=[$gӿ??ƣDےpF3!0W\}V |%oݒ15 ('sʨӽmRcgA^QZ Q/Zc)P GAT8b~5ZikFnzMA7.i6.x"Joaw^D悟v*uR؝R $;/R!vg6;m6O$ ~.H_uK~Fd@ F> "b8Sߚ2DО][s=' 3%~|_ _sqk+ј\Mno诹fC6jT?tP[l,03Y"K pmhXS>=.>/lrEPҫu.|Z`-L'bdU{"dI@&QpO)DMD$5,~p~ xy~VRK2 dtGU,zR֗bqA&TХ)μMw+y'Lk7FW]ȇ}kZT3m&ƛ(tZ4o&6f.C-k3ٻ޸n\W( i].6]-Pc7{f<~۱OXRwΌG<:"zHҸX /yp |f%k|L4_8XlwFm}䭁er*6{#eqQE 9ʗYȣ 5KX2>Z˨]Bgu!ůT#ĹEѲw3r c4ny$rEr oy~vۇhW9Bcu(J\!ŖNPZYWF疥.C)i =sM ѱ781hɫZ PRA\ ~56zH ~fʅacB! #bbɂ zJYQY)t"_OU.A-vYcAsc Fw֫ Xߍc);inVΙ(LTIU1h1 cuQ;XtpnC׾|Vk'BG̢8U&ÁD~B&ɅM\ Ii&%N.|BeGyXf7-f/giuEHE]H2 :BʅѿA߼H1$=e T1;9E<‚L%[^UUlfrZ8_ \)hNeވ$s͇,_C,kňUX~;\ߨv$:ݮ<.F|)ZrF #<ƪ:tq8<؟1Ê*4Tz@e h;h^K({Y=ސ#HoٛA :/e+фlj&hPA!UYj2IDb,"_KKEi)-IVBƑ3MidEK;X,/;c !%+&UnRc6d4ʣTK!Zn'}Ck bU\Œ\C)ۇ2,c濼j1C.wVUx$*duvle d"qZ,􊅧waw(wrPoree@v9 ҺYJI&d7]]-޽ml9ܯ'Ӟrp`Uww4ۏ+Cr[kԘe+ǟ$u@:?.wuMgau;#?5NODlQ볡,J] ݺJL ^}w>Ƶ>}RdY_G5K#XR VQ+v !+ΉTiiTCG66oXN1._`GsWR])qS-0}R"%M Ws`g%snM`Xk\nj8yv gw$xrwz'N|ʥhaaa~P9L=9Lԗ0uS0uS0uS0uS0uS0uS0uS0uS0uS0uS0uS0=j kd5qlZ?qM3N{ٸS682 _^Hzb)!LK%LE2$L `$L\)sP)sWG> 7B3왃uUB,"gzJD:HE EA-D`UL@.d#FQGfu ߛO4PT ز gN`RT(6jU֮V!]2ztwYh`P?x -Fjp=_ުASf=2oqNGAcacf>;xwǸI# *oGTLU5=jB(1Wr-B:&J[&xC("Di/>_`)ʠ'/y1;ϵ+lveWvN>ѯ?/>./WQ y4|`㶛y-ol>y'Vȸm|c)\;YPgAmSWl~J;z$HZ#8*+ޚ0XՁ@pӪ/M嵝99DkR+al)'JnGH]DƘeL{54k6lBrlQ6k*@Dꔼy0U|487B,_#Ű;U/Fjz~%}#گ^ѓe}ڤ< 5V'9n'9nZ:x%eg9FRV=@qoc#xGAlbPR2wl  !8L-sÄ0 %SUX49T13ѲTs$>sͶ8UtkZ:.W9ȠaT!+JC*і\ř:[<87vyJ'%2 g޾Rm_hݿ~hϺݳó{̩1ݽ~Z=:^/MHouE`)gWL(P P{ +|=mKp- )Ŕ1ZUT&Xcбzu- (G#gG#*v_#6CbJ h5@a,&p^Lq쌞ĹQ6:!8ylMKwe֣3T{L)ݴ*_^EmSiH.ِk](լXu걹x%^RN]sGAPAc2}uI0\EZ \p*Z88`pV èzF.̊'<:{u@[b0#hR&Vj)uBhgQU6\*Xc*CMlL0 :*0 cC} ;_{.:M9ˮ6Pl ̑dnuZt^1ȃ^NEF["wU05X1+d'7CQVي牌LřxqƆi*1 v/1z8g=Z5˴##G+ 'BB'<D~}B"r4q <\&-s9\g!4|Qr˽9oE/ Yٻ6lW=MMɵ ) A"dHʱ2~N5lIMnbtYSuriERdN`c$1Ȥf&^zK2^'gzQktkGU}S".hiM BVPEp%mPsf,eh5ZWzcIy?pMz7eLltG]_Xr<,ue[aaeC4So"c/vVz~1}h~;q}gw[P6W=~+0[U]_y#]o5ܟvak܊,@Ή8Ħoc$/5ҒxCVֲ̒lh'bRLzJ0!eɡVWqHUƹLme߽9}wL㯷Gdz%c7=nf[0>DncVp>ټP%3/tP[ dy꒐JkƄg9T, i r`xgU)kCdepnrrLkeP¡P@kW Fv ~g9.{^W=W9n)z,Y"P3TB)O&HO&HP+ٱ4T-Y lA*.. o_@@G  bJ̅~k{:(3 TA4] tϖ'n){DD05ҿd=tD;59\^f_Ai:]TpsaQ LȐ? Qp8^:sP12/8ֻwuj˩U)lC][L<ʍB c)/ךP;ݾ9j^L,brʎ}->yen5*OX`i4zBZZ44Q[3^VK<: 9=-E11 QPt @Q39)*cL˨jkjl׌QAta5x.4uuNuhM9YӬhqcvڳ1pr6'NMڹ N#"q%n.Qd2PTF[.r*HMtP n#* +M`U)'-oٮxWˣAZVڶ=T󊠍:A$j#( 1|((݄TcfIoI2CQ4#$fLD <AI%ŴH&t8h4R9%^v)i'-a;I'qsbybVV* " q]Ҕp$kD*E-7Fpԓ5Jh ԃ#RcUwy* aL*<k 9LGxIԊ=R2.tAX"RRf`*6 !Ƙr4BEvX < 9sҡƆ 3\']LßF>> `;>^ʇg/?Qj0eHPe$'([2 os,R(S9.3!AUT* "R2, He6j- Wqg$G<\ NzgeM <;2jr~T-23~_D*5 T(/".O~g7__y_'d/N6œ\]65ͨ׷.gA!ߏ/fErQζ_ 7X@3w )o qXܡLlhzST_^.! % #Ϻ5r Ş~c'0a֫)tQ/2aN=cw"+Y:t;xkcr{MG쮭"7?7%Fr+:C'ȴc{KՇ+.`cBP;p$E kSwMxKl&mgbc/杺PiX"j}}F.">l6<]('fIo@.19]fuum(X Zq2K <6!pEV(*7CFo>o&Ro\ Ea몣S()BOC̠eY}4a qe/e _fXQ ;&<_7 yRCS TFD>LQ^OUruBL`N&kd2Yg2*93L&Db%r&3 ~<-+`.>~X1')Ylv֦:{łp?!S^A{ NjwvSÞohW,j2j4]\b 6U0PjeH1CjO!ю#/cT)P/m&3R˾P1pZN$2 C8 -͘FH!)!l9 ^* CS7{˛E|څ 2 ڍ (*B²AS!UT8=K`2M]I8M9A# ftux2JX0^* k6]fE+=B2;pZ%Bym.&m| a)uR66a6L?keDžmԉ\bk zseG\`Gs2r&$vϻpUaM\H},{0Ξn,4;gMuCߞ'5<0 9enVt2A_d9&¹ҵޞВQ49@zhk@Iϓ+"H!7\_,_ Qǘnu~#_6ڼwn(vpe?-hKDp 9RO;+yGw A<^֖ɯW_~şPʺ\/ 8{WeFkvRrx_!ϗaEKRZ%U<" Ct\݋Pܵ16HķUCG{40ad#e/9 ! ]eLG>ASQHJqXȼe?p䥢OZ' I Hƿϑyp9j:h^ ~cQ&U"%}bg{ȃV&#C5:L)Te&;|qRTޖBtƒ515 @Hhj;&<0{80{fyJTz*;&Ed"$[0v_ͬ&y37ۓn`F,/#:И+`y"SQ+d)"7[UTޛX14ᴜIO5}NIg 1L.& -2Ȟǎ&i#wyo Zz;zwa܁])CXG;oq{VmHve9{BV}_v` /Dl;D2wƟGu1E{zG- ݸjKtLE<|s j^TZFg,vQ=:O!UMg؊3,O0ULsLSL%@[z@cSGOhXE9"S2$/e A-6È`. J%i%1JgUFR*u)7X T EĐ~ц\jlw֑nnIc*̶Gvam8}ڔ2]K-\}ԶHmCAL@Շp|H舺I0%xe/94Y1 ,F6Eֆ]`: ,d27 !:i d`#m$$Xؖ&h4+5&) %R#X΍WiC"~rTH#si~,ANXpe/um'Sk[Uq xQBʆl- F'V*)0 z!jW^Ie!rcCQ=(J:Fj1դNRU̱Rqi!'`І c!fLI2gee>FΎ|vdgE߆J%P J1/3) )f&]dq4d\9 L * z ׿0$2!! \ֹ౨pT%P NG ݧ/&&F CVrVd2@fIČV)|R,&t/8xQmM<`RN%R:*hQ zcѩPlTHIoʞyDR+)>5|Ֆ}VfePHC U4@@Ȃ3 ,D: cD=5a|q},Qm܋;n@ٟ;}gNЫiQ _h}GԤi4&ϡT $J" 69PIٕ"Fl:EzX[{vݠ9iO+DM␛rA㕷MJ›}\XO7[#=1,`TN(##[z!&霹";"XƵ֢9yc{ugAVz`խn}g㍘rDc}~I2c\Ojzz8^E!Ww/4 }[ۆFeq*Ѫj_`!5^"=$?+N3~[H 14bV6@Gr39d|IV`ruS,;lH$s%՝U-m՞3wSVZO< !%eڊn5S9@H HHKFZ$dlfjz_41 4=2zVί i'\} cՔq8bҦ] /?&?uwWb@~M~+6&;>OQa]=\%EKf(+ $/[>ku-rz98 -, EOM:< :7ɺ $/ -Nn/iV6̥s>mvb;uz~`wOf;˫h:i7Rf*N&_Kj{7 'Őz_ˮȶ;Ϗ:5vne @Ωf}&"ThQR{o٤.Eg&Yos(-̓QYɘ4WC>IQkHJʐcr$g%٩3V "9cd:0t.3[vFΎĻ\L_N4Ӝ}t 4nnwuy6kٞ_|^B;{ : k r,c9W.sI*mpHύsLX 3%x*{d9:N$dUplA`N# Z:bXjl8q1sW:ޓrd9)fd][, 7\> B4O&Hrڑj: B>H{ RA A*.;%6܁Yn-4dF8ˑg6 k{:`n2 >%}Oxe?,;'wZpg|ci+ApГå?<{~zBHQgh@ρ!*D)LyXAȽf@Fu"n,]FSL1x,DD!,rƴ2[mֆh@"|!餹 C #WT(#wl=iW}i===i{P$Xx>0q $7 \\K`H& GFY4A|S ۓQdԿ5u2ỵ #P(x%f%8MTqEsxabVfX7xA*ŅhRdY)_* \RlT,pCLlц,j;|k?ݕB Ǩ6>x3yC} U54gOκ??a`}}0R=R^|FȔ"M746lќ C%b{<Ǚ`s^yie0{hD"(D G%06~*TA!5)kۊc׵E>yod V~侐P oBVU[Ҕ-Mqk!D90.XƻdRUY "s0<A &޶CaaXdmYd}p= 73?hK04!HPCڔC1@^.1:)o]lq73Vr'|A3'Ŝp,4tD>X&سRtⷵ1,mW\&^R{拼Yem{ظӋAO!\'+wLVdݵ|uE<|'z,FOGw;].њ@;iA&y@[ޢ|q2)bƖq7_&=^쑽k/TZqMdB_oE_YL6  CQwӉI?@/.ZYIPT=kTQm6Isd|bP SM꛶J4_=rnkt3]; :¦UG,Rh%B^OA rGӿ{7-a^޴z~(` Cm{ϱ.\ǝt1'PW@ZSol8Rp&QUrN9.I &EQ w@i(Gfy AM!ykXJsTJX0F-lb"&=ĆQš K"fr)9ͼVy47o|{>V&mEۦ_dd ŝVc*%vճj`,ѶVA5>(BTLuhV,%^@8[L~~5[ҮY-d 8]šx[*H\%ӹxkPir.{K5 ]d&e/ vY&qe7:E^1v%k@ Z juZt5ISG\\P.Y)TH*A*,+FlljS\Rb8ۚx]=.oQ88cN); U f!brld:* o4پߖe}4-*_|F۠X,_ˍ[ThoC*q&'\@m"LV'`lwBޘW:7TO~+k2d yLSL19_44FPf@,d%:!&p0'CCE}*4MNCZDCG"8@rw7WeY%{Ң.[[ɺ8?Hju]C5u59]Sz5wz g.Ct]oHK%IJE,87<[.#[.ӵjg+0ڨw#ʮI!=&;>^|^ϜJ.Z!yt/z!f,ӧ<}QW?nHdK]c+unЯ%ZvY!}"{ Flׇ';S<.uT6EY<Վ([x)2=Ld ;}aZ} $WdU m{V!LC +  Q@Ri $whJmGR..zWmmR9eE1XNB.r "KUtbRС8[>oN1=e&5iumyg5 %)8ٽ_'tt t!j:tv-X#@Mr?"@!V5@$ͳd 2P_ON~}6ڴ me(vuu{t{R+ NH@H+3pBM`N&Cb&n< &-cϐhRF5eHAqك|˟ G{;D>Gzp`Ѐ\PWe5/yk^wt,IxD,I؉%}G3X .*)ĄC4V x5DYGTwՠŒ:YDL)Ty[9N5)4l|#nlْsn 7c݂=X |nP ?ń$3e9[p^xz?s@.n$&Z Ut\ʗx!t}O%U~=Gwlw]vm_{4 Yμ2rzY3 uŚ]/=놉9I}ƒl^SE4m8"}vg}wvI*peVy~֡m,FnVa۱`<ѹ?g-Th`k-U140"Èƫ)vb)v?ݤX\R %ZcꓩUb+.k[cЗcXdMRAc΄M&YSBʓ”cnlDۛ Pnvs:;x!}?٪Ud돻^v 3lbh+ѐU!h Y C >GMإ]͘S%jcLZD Kt0Ś`-I`#IM"z:ƞxdkE/Pr<"EA;$&=$x*I*M`=IIV]'$/t5;)??=Z'İh `Rflxʅiks95-~U8/Ytzvϖ+ we+I#ҋ=?;grr%- gt7_.}\JOt`T2eW:s*\r[\oUFE7.ɐHhM08zg F߳YMr<(({EM`$קȬS \lSY1*2dl9DKv9eAGcA"c5g(="WFLOF@T k)CTHlףn+㭶;BlNme5w?0WwbK$# BJAQ HRQhcAuT h%j1/Qlmb'Ld1jl,6JXmbk^ .ZoڱՎ՞*zRZ H9 RL>Cj'̍֒T!uHBEm%BdD%#oAF )sm$ hkg{Mp< .0E>hM^=FMq=UDV,X}L@Lh/^(H$ARК#ص9l;I&M;zM;i ہRt,(?+\O4&Aw$}[乔^]Yd\`aImd,1Cq"{mWgao<:bzYNj5.l\U/B¢"do8 0:ª:#yVȆ#HHd s6%f }[Z{Ý̯[aո}~r-@^v9 p[R%"-:<$24xLJt<=a#9-p\E]D0M-l jumn"5VE"M5"LԑCu`K&U 2g c:QoM Hlm;mFA֖E73h1=^m};s4~AA#B5X(M9X 1qvQz -w3c3[d2j'm7^S|s5F}B"pC^M\0BiФGOiФtHijq+L}R/.;lVc~8$CQk9ʌˎ 3c|O}uiDmծ{4#QAD=JcΤHZsaƬ5 tkaM\ʦ]a ә{:߿-_E^:Ϗ#ֳ~.;o_z*'?mQ)(BPɘEm9gVo%Hhs%LE؉Ehky'0J-Ϯc R G|XzvT\t&~X <ޱ}w|O7gmgmָD+sqoRQDv&YE(3B#$- \ N幹? }|4m;lEYEr (&@'S*6j?8N&4読 zp|>8yEzLXm` Pm5E b `8P6TKդݴ#Zph`hPE#Ye'섗zAO!?Ϯ H:?{Vqx)yȦ3 t/n/EGYZv=[>s͎{P|~_*S.$>2_;JY%Bdw?rldW L尙ɾlm78!½߫uWz=ԧD}HTZyhW~8^l&_&Duyg1YPh(7aA>끢Gu%z,ZmR,"!L0䤐0Na$R X5KJc$4y:J>VieG«A<)`eLJ6O%sJܾgNTgN3s\P1Տtc۬eV6CjoņMeD-u@+={@ )i\]](}nh+&O0f+Խhֈ 0'mR83+Y@ɼ#iZvj6r㨌s2ID5M'YKڐͤw,ً :#g4u76z3%P)E+@5}׷M֖"-s͵Wtʹ>gỨr8^g$8V(6X3 ) Y6gkP^n"jcH/u~Ф^peCpi6',).d"+%($NJA6/@qwvAUiv1 gu`R%D`+Ļ(~rֲ`5D6b_'ԯ%a$P:$<֬S,$].B̑W [zFꘑc!{r'c6%@hx $T4dB&ZOdX*f#e^F^?8.?XH8(OM-ʩ#j9VENw_dC\ UfD߽6ӟM.70喼g<0WjTCHde҅BjNZKdBGqtq09MFE+(|! Rg]%n&נ3 r0}lϮƇ*Y%_Ja1:ȃ0ӗ`gv#Y}NaQD"v*M뿩k^iPnѵSsXK(ծp+}LSؚ>}S٧كb傘p\FOӱrl Eƿ׍||S/wD ĺZ[:Z׌ZߌkY^,3=Gb|'mg:U%6:V\:4ڰT>U< }+UZl}Zp<}=HÓO/?{O|Ͽͻo?a3J,֑`$}'P@|M{U޼ip{ӼlhW.]nĭB;L 2fNR BRjN^qWlfp2H*Zv*.JGB1:z~Uc&-jg( 'XG*!Y((E)PRlu>"eȣI$]oCu(}m|kZ>Z<)HtҒ-,YD98%Z6ـQ)` hlM+isE}wl칷Ҡmy].҃>LKD/;2i*ouħQL@�~pz|6-y wL#eWF $0Z~-F<֥}Ur󎼹|6ulrh#5Y!lcxsWB7]<Q<tO켘!PB%2u&AkңE:4%6)sNEMG2e)J&BF&(QMd`3rnsx4L~C!oѳnH(9G9VJ^*2G3(yUNErW ASOAi2:0hYTF z^?hr ; o@WMtD"REQ ?XWJQdBפ|0'dE[)% 9 BX$ 0cLPlPV(6=G~kpFy17n2;A`R,́gmL8?3Hte3!;5|x'#',n]ܦik&7=C] f>6g!R =lueL>ON:9M,Fw8o Sˆ$O7DeTL#aZ%]p8b&/Ѻ1,$hNOF v ZJ1j@xga! > VbT;ve{9>-\vvF41(e3zI @R>zSqN q ]&"Ƶ]Vv79,ohIQn q+yg\ /,92>y}{><>2^Pip'?Ǘ0<[zu-wD&*uum,Ԓn15 btsjQk xnO1qsQ݅Z#}\TwREJ;R(s 0'մ8z?t܁~hY!4(k2U7xl<˳;,G^Yv/<ek`4I"-QZfڔTuiYK:d A,Jw휱t{6F*MHʢ]q@aDmIF1_t}Ig~4;tCSц'M_;EsGc[M|˽"/2^#~X:mbByc2]_"zfBj $;J,I &(Jb08b9U" P 0V`b.= 2d _]kj4qd1Sp>:}9ӄ?rl7fl7[X@"ӺCfƪU(N p9J.; K PD5H^).-S:J+rn7'=t,HdWj/Bk JK;xN?<؊5VxUm4>E>{#$fhbhS[$e {sF5սqpg/udkcR4CF_Yɬ-y"fiQς(A;G"ࢲ,P99/MI Q 3r=v6g&ڰPL;eabr:dA&RL\ʗ 4'ճ@,sYz}׿rf*t& ;~"ؖDzghJ(VgP+z=דb4.9cMrƉ$(fB@a taЙ\Q$#2`csv&Ir`o#Q$ cb!D,\YLB:[ &Q@5RmYD֙4<ѓ}aΖ}w|,[X YY8JZpx( kpQK,D̸ سGwcM:>l܋;VӖ1# F/p|sE\i4"!u $!X}$U]9͜g^Kql&;lzY=.ߣ.=LI9g- u^>C&R3=:]rcLkW3QAͰpLZNXsA]Cf}߬ߗ\#a[3UEz0 ft1J f2 $B[P(r0b{L6J)Dg% !7s8@D/] R*kez[r*(MWG) F(G&-b5IUDɈRW_Tckglz7S̾r|vzѦ[c2/ɺO.6Cz@|nu.x8\L ռY -UϤUu5˹t["769N6'UZ?T*[ n_K7қOntQvoV%.Vnk5_ٻ6r$W}(o$ݞ>af3 Yäl=}|HԃX#ܖI*%r+<&WWvst̻yvۉ;8-Α.. Wz\>鮻0O_tۼg8k'O tj6G͝5}Ћ/붸/_B؝AΔB!uфe T}k\ Z)<cR(X U"du&'mm2*윩JB[ -!e3>h\*8$dP@EƱ{3sv')bw=kߙzriaǞ=lvcO4s?޼Y/l/?OoZ1Lڣ*#ĂZ%MЈH:IsfF]$n&iK pXl@bJ" BQzE#;UA:9=I e]Vw'_C)r|:!-^tcĔ {#rPM쭍9p_}N66j{0?,$g7E)]UVKKXjHGf~zϹ^G@~g9e-$൳$YSU7\l4t\@DoRmOL1j7)CDl'LuIX@ m>s@ :iev*t̶ЍߠRy`ƚZE!gMӳI|NsRSZ['ǁxvɯ-Ay$=-S ֩318Htuf%2W&~ ɫ7|Sȡ?AؿI}Ͱ{zmi~'E_hǓĴ4Щx6rS9_'s*ZkTOW+Nr?rIJ3)@gH.Z2UdamMC1Cw=? z'˔hDj<>t3GyN jRBD3e$*90VP6AԢRX<m+l2 ԧ퓥ӚE)e͂׿6;>?lz*[_KjEP[S2*a] Dκgd0L"v9A:x  GZp\5fV4•!Km2.[$K%ضPjixc<en8ZYvN.I5QC3svȲ`o꫅닾>4e>wtuV/`inzja 6<5s[r7HBQ6 kd Z_5]U ET0 O乚ܼX =6q&ᬲ$0ZkE%Je4)F2^a%Hz@&Szqgzo'p\%,UwtaE(DY1㋔e~.1x}3[Y~ާk|vWт8T2ͨo"E#e1ݍlėZxuEǗ/_9'K9T)Ex!麪-4kVJX(ayk6I|\T?lq2y#xy~3} /fʃMXL9E[WoݸVt963-ؤDRk` ک6y7]\&M4//c-\#Ոk嗋7J:E_D_?~-HgOO9 :=!dD8/HZ]E8 "ߡ"{s'?>M~H b6*h(G]*vN ' 61 ]}\j+ Jd=xU*ʦ#Mʲh#Tk ,3&Nkn\B络w/^:߾ڗKz_͜.x Z!䀺PI2kMwl$ <6ym /c$Qrqc1̓nqk>1NH)A79vO*0>[>9(},N4qe`}>}V&, ̄2d:S$!9")PS‡dImh>h %!rIqVeJ٘53g;iٸ>ZTz[-9G~̝08ue_o/ĸu_sw *dɡ$KJ%-8A@1E&9cEAo{ZG^F۔Y5et&92),9ܺ8r3svԑ{LA){J1F4ϔ.Z!?dme[M%c-刭nH"Z %Rs q=H$ʢ]Xe(b1Ձej60Kf>XB<|bl%D yBT$ 0̜.).*#у&uj`~km WYm ArA[gPrᰖ1hE=)uE ޚ;-e^G]X|}ul CɡʓPjіi<>RkҁRR)%Wm8jSIxւ{{2 U'DŽw+e0_rW珺63fLƟA|R9ouB̥3ΕPH,CC|֚Kkq=JkqZ!!2&I%HNjƄR d"ZK&Z@{o]z86#tw_Ip]rߠ:ҩ9]qw]8RW!.wfe61vNGЉDKHޑ!9s{3Cxf RZ(dE"0-y&h)<LB Qb%@)A6W9XA$ɔҁ@9M46'ࢉ.X%(;]fR4o˯@zIV羄כPr@B;$T'UmAUfzܱ%N&]"?RViSyj O?qs5yI{*Yqv'":Rl.De:H>?ry#hw΍:}Sin<(7FſQ79? fJ<Ǐ`{SśSr];!狅 ˡ[GF }[4}Sq9Θ7?p\ߍdThCψG[FWe4fT)Jӽ ohF(觚Kvo7Ǚx=z]_ oFㇷvxd`9~<obk <RlŊi`!saY.w~fn M=Ggo']cK{k;| UI(f7t{s?E#N([dkRS֬Jlj2Cwٻ6\WXysdѮʃ=IeԾW1E*"zx%/#[Uę @hx.w/sԮǎty͛l!4NYlܣdZ+/X=PT;ݧL}{hj> 2WA`w{T!3&QVT GH t^bZyūxGخY77SL4ыIHƨcڔH K`[zfcF'adN rpт MuI<9b*@c^" c퍷7مelEg3sJ{;K""7 ;Hu-i%P[\׌XߌkY&wĈ1 G,Xvpe>6ɦMCW*V\꺱V%:y4|>7T"UJ)?4 6TM>}Y8 "ǿw~:~/O>pûҮ7I h]= <(훶y%W]z6[+yC^ 15ha $ޯ7KltyU0k.aWӳfh{W2NIHOXnCmaclAG1: 3l3Y 8Y ǬG`:jME|@Jq˼ez{ÓruҀ>$€.sd\tx>ڡjtr3Nw-Άh⎳mj{vĝv=U9y(䷃^իjz勤W(`X!sbW9^1+xUQ4j(~5>\u=uўn5kڸg-R6DvFgZ`pI)KmguY0k]AfM׎:y&hEXZms[V mby)u{]F&zCj;7;Km @|EoT \Yan4i+w]vbNusRl/Ԋ*)]~W%^CZ=Abj"Wt_@y´`$(p7'w1R~βi "S2AyE7 bo* "[i8;œ-i}V:PT@:PTճήUVElaulK#;[D`C!Z̬ŵY-fVbfY-fVbfi^FTE 1"W\+,cE &1"W1"W\+bpE 1"W\+bpE 1"W\k 0YR9RJA~5)BJ~-)Bx)BKHG5EI u 7Q\@t }LM(8͹ L#ģ˫#ʫj^]!SJIIZj̺hg;$q&h4ԠטL'SX#XRƼN/]uq#IC5;_9]P辚o/*N^pL +,hbX3[$keB4%`І c!fL$c>;[ّf8 JŧQ%YL3.fǸO2Yy &lv Cm:e^ r_dEύ̐ďP.\%H3SUT0T;I?$Ǘsk#Sny!+ 2e 3q$JbFF>)s NQtF$g oM# ;h3!WAdN:7 f AU;$_ -X3ٗl̷yBi9$MdLp %"" J. 而+{t e /Q^q7pXsތ_D4#א>*%dPM^(؄GhNg[fנ9y;LQ9*Ǜ!7ֱ& %|MZjEMXෙ>3A/;ÀW ,2Z#QFG66BL9sEvwDk%E#s\_ 1zGҬRCԻMmAlO}SCOi؛se;D{BLt6Mr+uK?]߇O5W*^;u:'r[mxb~O󝃊@m)'%%}M7NR ^F{O7mWUT⟼EX'~߻[Hm̀ΖێuO8z&{ogG飑H8;.{4 !cq--Od8N¼0nݕe`v-ؗA٭J FV4^X 4t{ף#5lS\JR׭ `7kb[|- ͪ)DuHK#J#-MxiS(+-9ddgH4NƤt{%@x!eѓ:IrX Bg.Dr(t`\fヨκew# Z܆ ӫ_N<v;؆f ye\Ly]cxҕ-Q/5Z`ɝQ%+03sqi8]R+)U-sU'399n*h鈩r`]%w֝9x z4uG{_-}5S56E`WT&e [ͤ?j F?X@4rB&tZՔ ړwC~.muY؅*a;\_ A20/:.zy4 ; K)Ň%4K:g?h{&f7RgKw4X/?-元MI4 YlQDNx]]; XY5Y#e(&կ.F0r  ; Wi Mmܜ1^t+9,)Yn-4d˳AŢ$0iWvEre-.OJָv!-{gC9ҧ\6ӆ<W7'/eRr)i2ȋ>K7$.2њeetDy{wZg8BH-8n 8JBRҊDZ3|DUR0'*ȥ1#RhK1yש#13VZ\Sd(o_19oJ"^Eo б#lr^"6kBu p-dtF'{fB4VLu;/A䤭IUQ1fYɁЈ1K.HPb.L̖:@93)j\$ug Qƾ;XeE&7,h f{x8̎WO OKl2H0sAK!ĕBfT:IltȈS!H2MC\Lz/s$66A$'$(dhRfbl؝uqGød\u+mRV]5MoǬ#d.H r!87"-&ƨAEEmqLEd!fȁ !YCH&őd]p>f@Fu"nXv֝akԗ*"`<})EhUX%F(a ɱ` wmH3ia 2I2AkP6'662Mͥm3:z,tĸĴ "i)K0-%iw$%gidZŒ*͙61mwGn=)47خCH723ߊxŰ{z}Ngw43xs͎#9R2&H;K22{`@>KH1}}} ZŧS",HY . M]!8O^KJ?u:TKc5S/,]HBB뻷o^,_CmpKH igEQFXSHJ-5Dqw1గ$8-)j.xwբWsƓWzI'ݾUz]|׵Sx'>}Ն+@py3Q) AW XeSHAjndoCAn/ZQ phq 7!S傪DVJ""E3x?f5`ܤ;w.%̣YmVs<.`B2Xx&L,{r,)cJKrP0d5, Dt%NF7|o< ,έGq1d}(̴PMfn.q*Zц=wc< pOCwCZl*B @ q1 + xlz?xpCgtv}o“j.A\#u@a>I='+ԡ$i0$F9 !"9(tA:]0EA3*@!t^^%G>T|D `+HTTENr:Z Ry$,FXGiT@FTsL(QmGsudXĴǸ#aB)"VB"IӔJMy m =MZb#eB\F}r3OS{$OMP1sLg/Z||=9B!B: -,% }A9[qU98g c: $ ,28RK!*X (mYRAeq7\Gm#~Eq͓I)CYevN>2 8eQyOmfX3m%Il"\ lax] W $t>WPg&# 0xL%LxQsYHZoXn;ǫ+4$ηI蝥▗VxbsfjR|U, 7P& ^ 4ت4).)RP;xSMP>OU6<`tYizyjG01@TrM(h00kL`%&ly< Il;,-k=x,+"I\LF@i鰕4Oh&#`]Jita=+ e+tv[pH`hPEM{vmY:|{!KMC77Re3ޔ_#Ķ&m]G1fuL2G{iz r6h9 ssK?UT޳H[MMق#iRư4a1H^KP```y%9nY1 o$ˎy1w ZvƐQԗ'4 .{[u\ RJ캙ߓ+KbY0 0 &q59fA?@4I" eh3Nf̂Y03 ff,mxjI3`AY03 ff,3`f̂Y03 |@;Q 0a`vNTV}߉JRJw(~D5f"LP3jN|5f"l2j&B f"LP3j<5f"LP3j&BD5f"'KKc#doBgPR ccǫvgԪԯ5RՊ]N(^rzZnU*g9S]OZU{#Be.ۺQ\~ ~9n0eH gzٛEIҿg7N,釗oO+Bh&uL/F=Lnw_Bjxri 5< z_ؤi˾;w;B]_? Of;}.8[/7?N0lwELqtbsy5#zNj#*|lwo6's#!]}jTT_N`yk F'1:#wR_4) 384+OY_U*Tu μ Wflޙ;yg6ټ3wflޙ;yg6ټ3wflޙCaGAh#nn@Z !(QX|sp 8#S>|dwȺ!ʻǒ\1W3OVf 1>Re {AV 5ZŨ Q!cFY'[B7zf!% 3Bs46 qg.?~tٯL{1fo} $y%3 g*׃y3f ixQ*G S8^-)%(bL+hbc.f`Bq VQ$] Zҡא#*<"t}׆yf?VY3ѭnݝжrGRDRQ""`Y2c\Y`(b7P"\JlB Oc6a*5 `/<wkYh1q侘zEؙ`$I0Bp2:Zd=_^E Ηnm..[Z)ů`UvI%" ol yÔg*Jc  uw~)oW^΂%€0s;qR=ۙ"j߮!<7IvB9ChUM).rU5du5RfvE+m`4뺣:ۗѺ)VZ7rU ŧӎFu$UaTWlc|[ ;*`L>.nv>y_^߿8ׯ~9})&/OfAh\,»PdnT߶Z릪ꪩbdߢ^lzҡmZ*[]}~Czr'm:Mg>XHA[ j~9>`=lRbBTҗ&Dh(pX#ͣe=I8JL!GsNE"R V1$iŝ4;JPۼT/! .Hrxpm2㍔+p(Kκ^e46&!0{[OllyL_Zht _Zr 0[f`-Lrv(&is *[-[k,GӌO#%vF#+M[ZRLQ$MR??@B2@t/ t _TfZSѷAvsr2Te#*Asy5Cy5)*SX Ya*&R<Noo_D"%T9vmy|$)fe tFd f6d6j7rQޠQ:6S}b bTȚX [衘ZảJN~ߪ )s/5N{mFjl}ʙ(H1FkzݳKE8&RsG-*T,Y19Wr& ћn|5&dEOU%%kqJGX(.6A*v~׾T *nж.-nvtzopx"S|a7 5ƒ%DqlZ)ơbʃwwLHnp{vݒ,;ٸCvćLs d4݂=vIQ/ T#R>vmwN*ۖi. Z W/ 'R|<3@x?|~WNΥ]sJ>8iguK?_NޟL> hỤ\B]P+|(S7yMDR=Je[>4=M@KK@HaBZ!a@E-'y9n1ǟ줃}3>:u`nK h=RXb΅ k;Q. U F(EM [,[DC)sVFm &$CɪJU4jP%9U:딮oDFg ޥ±[5-׺M&rkjg{z/ ^U>F,b@=7ϝ%NIѸ!jILױΰsj!ˌk*I*b h]UJbӊD*-aA-w.eK# ͟Zupb^":,\4)X'!?)Eݚ`RhmȗZ+[DMiԣ8kjB^d)(m4>UU>{@Xzm4rfոRvP%ڎ|Lqnr hu!)ԪR-3Bֶx"gViEߕςA;(ߙ. knjʲ@BB&;MXpY)+83.я=.!MN wt!⎫%王3'i)G'7w$MYa%)-d07/NbFo3 (a83Td2ZTO0cU%w83 "6к&v{H*zpǬjMݬERJ"+ S-T `U9{WA u-L>|dЍ /c.}ݸ&‡م^_oǷc0LvsWoH_ uz^> ;f}6sw!fz!7zrz>sm\flJ䭤M'}? mwz}BLDzֳütLYϷCi9Ha[7ww0V6yl2 /Ǽɞm{K{;:Aa?Aʨo"ؚ/ntsc6Oyz7zJ|&Wf%沛m;uv' |0dĆ=/mvU/ERJrE_S̺V(xv]йX0\Hu4'BGR8gׄz8+Plw#gC,4M|} h"ߜ}Zg;n4CWݼ͊^UB"D=)TVۭT(Xt6g $|5L`j-z@Hś8 !duIҍ9F=q3F6fbnJ+Fy,K>os&\Spw= V;Ǭ@i=S bnW4_s5\QOpuER?ڹNUb Nޝp?x?Vr!K ĚdO6)kT&(lb"dC%yGm+LQ}* r B76Sz"gbgt@bk<$VrwWo? pw vMI9E8S(UfeMԑcqUM XŦ銌]͹ƨZC˜)ȹ\ ƵRp%[: U@72v#nd]Ky,3 yG+apͭFBMNɟ7#Ns8N&BZD# S1&Vh2UJ8[ؚu%fJLjB.YVp}ZH59s7bI^1.wF;ھ3jG ڤ$%tH:'\J&Z~M&w$6/.$1BH褋K r8 FɃsnLbTx؍pfO>\0 "vDD|GDaNTD_JcRKYgSud_troXgP׉hHցΎ69&b $4>:茈ȹ?+QpqL%_g]P\θ#.xێֆYC }⤍R R~}a GZ35caP}mܴEFd"<3Lߏ\|G~G%n4i/=y xjp˅ِkbےOkj= &'͹oF] L(}b_V #efLv懃Sk!!/tX?G.@2F푴m3mBUҶQ~-1L> mh k 4X4>[ʹAmR2"ΩPd%y+n{Qd?:Iﯮ/Z~-ȫYOgak}+:Lj=5=Xu/_,ngs cl/CroĒz^^!Vg*7 Ͽ %X@-NehE~ɡ EDd煇k|۸GEo<U7|2* | 21vaښȍ_QuDDOg֨&PVuֶV4:@,]cCշ|R)9K_”-1~)7#yzCF,'I"~J{RRR d Z[`Az\[CacqN2`!o tD.!u 6fe^N3;̗8qNkڲ8 ZD?35#d׏1ڄ\J`_[|[mu5pbH0&%$|)y/HYBWB(^QdBN:,*W2=Fc1Un"VJ 1BNFX$ b1]&yfTP@y{9|J7oˋ]=1;̸S?+$pEmSH }sH Έ')t`mP(!e׃c.C1':'Q (ˈ!Mw6%U Z0:d A,34.Jk(DW-Px2I$􈘊r ]T뼐ȹX%Iڰvxvաs7x^Fz:|&Yk1OϦ9L´aYLVD.&ʝp B@r3Lwp83cH ZA8* dUJb+BXWu%T諭`{(I,MI"y[dd@ XoluiE9w̅\>$tyO= awIPz-XEoO Ur|1@P톪 j]>^5;^L~%W_]^4zR 0DJJ=4FqAиFq6k]:(ZK9"VNKuk)K E7&Z"6!@`Nڤ"TH1qF ,d bdHZCSh:Ÿ)|5Znx ʁ!`y HJBZWO)uD c:nAhrsFt2`GWc'HR`F M~ Z~G]R6GCffg[e[nU 7YJNhcG Ba?b-6P'NjN)g$]ʵhMry5HР-Z 00RcFOPjUgyƘM 1$S}}2!C_dX*~se`w@hWY\$[j~ɣxu6sSrHhUQKbIreaIj ӋUey7uَ{yT;/G)\?~jKSYNa:8>Xw{hU2$!x#YrI8tfEn 9KUDW;@m#Yu$0G:6>[Y_,o3'Xwxt͘gg]uT%v|mn{V/s3ҒG_'}N)˯rTNbW˖e n"joH_~8l:OXgV3]s~lO:Q^~f6?_I.z9U`]Qs/B16p2t|;uɃ")P8:R Ț@IAN l#&Œt B(=a)}sV'bm;%/dc"g!KBE?M6`T(ji8ir3šTm&U1+?Cw^G^eqcaVX+7W,6lFQh\I ٗ_icqyp3"f̐{πb ኈ1Zm e1hb#x!%ԤOE,: )ej-#a:)XbE ,c?)sȹUz@< iP͞B[״(/,[UƇUm4>E>{#$fhb菚Rv2csͬmƹ˚qg':ZM`e>KY k>%5ֳd֬[1Q|D )2IuBP99/MI UHZY3rݟ͖ o(g7 !d$ 4!bPN)>e/r׿r,Y3UQ:cH^l(jna$:J jhГ= ?,ւ4.gaHKBemV@Y"Q AAp4uXlMò`MHCK !deAd`Bd0]k "´U7  5["e>ߘV^#{X _B.5/%a !jHh+0֥pGr/]&ܱ3^Mƚ$7wMIhPz QN 39me"l:ѲT:4 uh"@I`|w?) # ]4wElօr J5}PZ1 8j֙:ֹu21O"  i.y(-R[W+(2Ujt;SyOk-dݭRj1m녭LP*k_W8ȆA̲3,c8:d:2I0#UcOATp&Hm3R뒄Vʔ D"N^ePC#P)EiZ0Q (QzSՈ$U9{W$#JJ ^_ hQfS9͇Tu7]^[nj% +[vV{6.^oߣU->w; }b.דYV#\%r'iZ꡷v]^?T*Zww9;g[vܲ[5_|yp9oywty֐d#syrvjS8_X-/{[nnV>^i[5w( Q_HknnLn{ ٵqo;c}RB>t/R'㢮z)j!j}@}IU$gv]Й4QB1G TZИ@90h1S`9eTުjMC(7ݛ~C..'_w˳D|vl?4inp5{?oٞ_|Z!=cMlVރP2Lf y@:'҅ 3g$וO^1_I}L$oRK$ &tN)T%ȹ7WU}/V^, .qN1W%f<?os&zڝ:oP֠Ļʢt:ƙu nP(MRCNubw;|s|vn}-u^ '?G-E35OP&%">ѤJh`j@WS5@8ji)Od jWS)mjKi9ٌ=NkQ,]~qqԭ=zn`_i0x~Rnȱtxޜ!"&1Cv޸—-ECTɤ Y@NT2Zt%aQ1-JDf5,lAJj-sd܍RE6[X Ur+]UK|;{zBMiqȲy 0M>ΦW.Yqb0)%"Z?3NJ4Qόd s*$ EF]!+cTLMm]- 3 x[ %v3rq2k VR7~ޕ#_vfy\l&,v ,it$q؇ےle˞D&l*`WnVE:jF#FXIY,PgZ 5{xPB"f!exю X D9cH'"c}mtTֆ1qaԧ0 "6?6DD0 A-"]\9 VANa'&1 I&$@͆iPQzHL8cu4(&LR#I,( #bcGį>\[hvi^).qs0^+@D;(G3NI[~uPB{\PJI+pZ\. 6;6C0p > vԭk!)HN4|J1^͞| ? /I*p'ſϞi$Ϟ;op):ppN`!.>h)`2 W;lo_NZPJJ,7A/  Y*%zV LD(B۞~OVappxm0$³4JwH#68R_ΦF<҃lX@UƦY^]4Ǜ`_RsS𑥌ѩ6M׳P2.j[;8=9sf9ACH0n9I߼& L #)6gF |bwabGYCؠ5un5ܟpx`ZZci1" o;76E- WDX29Z,PN=Dq}{жĎщGEPOJ8w۶v,۠(:$P΁m'aeQ0a(ؗ&Aӆ1b&Jxfm*nbPڥYH,dݎ]V%Aty;_=A| 1~e҇`?.΅!8x\KB q1Q[+4FfF1@ ޵DH}-0psZsQ4D){U:H#5 Sf hEԼY`HPxdc]X{))᠜i;Ŝ(aHИ8X޶p5;ˏH,ywzJ')%$&{Y.Q:N*`Ez_ԧ%79x07P& ^ 4تLL"DA ESO3O} =KEdž<'EŎOSr()6$.{h< O2׍~raK<pQI$.&j@#Xi鰕4Oh&#`^JVzKuѴX׵?01NEM{v^ ِ']T>rWA(e1Q 0Ele8*<@pe@bιl 7xP\z#xc N[ jʻLpK=x=WgP >ae/&8 1ܹ놼[wfy݋ nsn2~ zC4 }T6ڗE_z z 4dO0*ˍ\пu-F+2gnYYQ:GqH)پѤjdWăWIg4Lߡ`E!ܼYҷ=Ti1]O߿ |/ܲKqsu'F}w/~~BRaxci?p\ITg4㏷en}N˳}o_ӥCaPO?wzJ|z7wƭPgixt0ǐb9fw/^'mUq|"d9ٜ~ ?T'T)ӶY{~An,tdLYbGBac엿O3Z@ H\(ϟuv^9ݮ4W3+S6}7E<=)bžAͪM<~ 4_tGіi*l2 "~7^xh~xKBލY󐢀U3C2:u`ivZe}9PVnKEl}%HUS;uidθ<7ݢ>\!sv JM-j9 8‘ubk^]%[_؞z/atMeR82UM rbdwAUG I9 {fOOƧՍ€0g[ӲmDt体7mn=c-+I%% J:[V Y^K,!FK4{G2;g)TVZWrY gҊFu$̡u",͈=%K8Vd*OQf镏3Bwۧwǯ_z:{_o?9~u7a88Ex+ߗ_h*ʫ/bdF7Y%^>rąCqӱP J'ߞȥ8XINQ\ i9|W*23\%JB*&J|! l z/+}Ek(%D#E9 ]$"E mC Vht35|Y384XuAȕSƃoGoX R ,Q:څFvvZ>x`Ύ/5ʒ NKWUm ӽ0 BaͳLOZrgZq_qAf3O/o3đݖ_AI~zئ-9 'DR SۤKTkU 0V}d5ߴ>&>Du$ &[Y;v&H X o8Aߥ{u6 fP^K8$]<~AR(7Vs:tӢ; %sA39$M_\SB`!6Xs-\I~k2ƍ {&Su=n\ut^Z-J|&oTũ-\1>^RGՔs:2 NkMv#Ķ֜Sjk\Ԧf1ÂJFF(IH lB~{е_vq=h* x0h )h-zEA"[5 Ơ.g Ғ9tƠ.1d "A>s9f#}2f/#$D̨.Nڈ3^SR *(*A] MtBO!r^P]zpq,齗Io?4&h;'5AM4ͦUN!\lɘS0TZ!%D\)}Xw|,TNN;>-G>хpXRn[nw%14;6[|e/I9vy !{9F͍kɩZgQjvIaLJRgNjM m,M1*!6.\">k-z] &__?x!%b 69uky})yM05iͺWkl┡m))st1d8&zpMβGWe36!2gcrLcfM>ptl \#˾i+&m:e}U燢|Öc/ V+bP"`%8`}GY`3[Q,Zg(Ό/ȓzS+0ُCюa+Qqaqu_,e7Oާ!j1Dir SnܞrآV*l~zR/YUoeJb[Fԇ> Lq"hT5TKA5p2B QҺva֡Sz eYл7-y"TlAs *:ݚqwƠ 9{T2׵+ m~`~O'KlĔ3[ߗ9k쟫[J,a~z4uᄀEKs.FX]Uz\H!j*:~U} S ލZGIxg6Ҕh-8!ޒ dy]lm̀ZnZCr  E)F&gqG7RhzZ_m- Ng3w+VoR_UMa˷Vtnr't'td^Qh8)q~5L]ZHK 4E, `L"z#Hxl>4-b%R± ~rTB֫DN9EwگYvW/yrrʸ*onc0h GZgis1ʱ'R.X:g"䉜sюASzW8vhKk[Z;})y }Jov}E97k%׀5'6֒DTrL=K ۊ-`cy h0qvl|͇tZyZ080_:;ɥqx|?VX=NO| ^O֔_Ϫ 5vq͎WKv/}zvڬ]2]wW!76?͟"X6__߉[xdыj9zv[.z5f{ƻT2RVt{_w򍑛]~9zλY-: X?2nT7f!-ji^Nr_zKh~pƺn:tO6{mж «9Rx-]Z>]ʠC_6%LJc(yq_daɭmU:$1DU3; 4BSPCCu>FcѾZn5S 4n[іJ%-r B-gamp*5"P3"Bl)k088LC)f%[o!(۫8ΏyڋnܪCr̍K:Ͼ/F9;$T `rV7]{3X#5Q>ze3UƵ) !#rZ&5F=d(Uz4K0qk8 q6uWZ]]zb )=sJ,K&,NJ\}Ssg覐*ZALwh yJUJPC(Y;@lL DrIF9 7 P lRO[>DMMeYqF#\,%8CT`JHغt> kl3X"!&QYf4\IkΣGE9z~7 GGrLa?7.x/Hp`Vy{DMj.޻ ;1j.%8,>8mɥx;~ VطRZJsiր/-jF_ m2*lᰕX[ȃm!OYm᭴/ipLy^T<99vr/<6G>fq ,РxV/F/TLF1&ژũ0qÅS Xa=Y"nl  dQj=w$ܼ臠ލs l&s':&4кI%eJsO-0q[o<$vq~YwHav f]ܴ Q,)!diЁr TPl&9\5Ր"JōsCVc! 8{ &l 8S=e?18e7imY]oGߘbA{9ߴ=Ӌ"*!^(#/;k wv{ g]e OfnDg.b*:~tte0L)Oa/ÿtbm<rs$!"u5h.wZi'> & ER9ֺm\BCC3ˠ̹AJ^~wF(ߵBof6N\ee@&g._-s6A>@w旇 n0q/ä%ä__G)Q㓅Khg%~ڌԗYy'7jirE>yǪs^18?;iXi^=o[-k1F&Ŷ81]dߏ柎䷕yyWjU4+oAiSR *( ;F-^y)+ֳV]..~L Ch-wA_C[MRhU ^^3=\f/V @h0,j&3l9WrijŔnTD`dsGg\t2>d -yt\l8bb{[hk;_yU]fS:su];LԷF~c[s{ Xm)x2m)'P$,jlm.M;DyW7Ғz5bCPMݷP>d)佟N 4e>^|i?T]DU]s<8W0W{B3K#mH/,LClp6A`6G5EjIʎo$%E$ʞșfwMO]x4wQ}6N(LeAȼƓÔh2F I(.i'|KKNUnc5*УJه~Yu%&3{x_:Ă/$߇ @&u9 RfL^9͂ZV$d8½p5wc R9e=.Y"F+e$Qɢ) e H9Zq }X7Wq|8٠8O|w(fxe< a`9CpV$`iހ:9PD`xFIXJq(,QI?29>iZQrTrwf!:idu݅AŌڎ-c9<գłfq& RNP@B'JQk!շFszG+jf7?P2tPRڶfq`[ْ\>MFLr6n@)CQ J"u S:)d`)>2|oƷ*]+p4青:J9ULڨh"Zy>| , M 5`5_^nU 0h.rxuL:yGpЫVG\}U]g4T+uыɝEz.~7G(^ a{JpMoۀyAd2ӳAY |999<갘eHJ2BR9OXrPlyeC9=?;(vikG^IEȅBOiZ~r(k߯08\O(˾ݾ}y?Q.d={W8;!j&}^?֘_m ioѠ.~<ǗA,m]EIOnqDV!~8<&.M` 8L)>0ȗⓛ4}rJ6!lQ6 G@.| ii3}aۄ.%|7ה:~⍲׬몰9zZ]s1d Dsaͳ9/'mlXݺ{ޔ]A³pL܌xThb"ك;pkj<Ɲv D`C^37(&\JNn bdMFVJjaZA vvk';jԖ3Z0n7,j8G ҵ6 5$jňK@.zJ$vi0!]iinl T :Z wW=#W \z #sjP#'S:+SAUfkPg(MW[AMPv^mYY3^3ro?~N^#Ul:ڤ@}mVx$Nϊųo_޹8\rZXzz&'oGW.|5coRUx\4ӿ>mYnj ^¡6A|M&KlzhTu,j^%;ZzoS&l+Df%_]>Zuȁ4;WJT`Y]d1n[yP!, 2g +GUdoOP80V=X$$unɇk*- KC$Ec#dHі=Xsv1 nZ~'^+77]DWLoͦKpl y%sQեQ8)2ٕҎa1Kn7%(4LjMeDk3Axx%LZMA4EKb-[gâLU]S8shnBv8~KeI& QNEKM=,Q0/JCmbu'ZkGS.THrr%.HVH@tU"F/4%AJ#푷Bx@#4D0dèEI9CE$xGc;Dcc?`gtg|8ԝ>)!C<ڒm5N vҞlwlMMvֵ+\a<QlSȽfE[xH߆GΘiMh,@{Ђ୍*AdTS9oJﶵ-.'r,Lj椽6ߟB3yjb0 ?N`XLq\9c3BAW\9U9WpEHV}YR<]F3)%U[Tt@AZG:$wX9jnzcsPDGPNPR܁pAJ> \kl7KInD4M$( Sk yJN+h uȡ{8W73FOz9#?M5+ѩb ɹxB`eLlW۟dcJAH 3Řh7rm#he"'=@O0z>@@ڙd7ΙvbfTDEd` 6!q41qJ4$ffNAFP$K֤dI*r0(ZF:)PyJ3.> 'dSV2QԲY8*WPedzzت sa(1GNUzh: óβnò- Ȟ4G19aVɑCBD٤;"+KVH,X 'ȈnF-r׭JbCƗ<@ۓڏo"3)ezĆ\+& AeBڦ~Er{+x(3YzU81LYMYfŬkŖ AP'=ϩ%IL 1hS8ݹtC!貫"f-+{sDM#[%͉ݹ'VtArI`=otbLƕRb4\\[ܬeC!57e`eU~})*WBV·b\!YKJB43̻)DM!nΑ>r[j5\髣W HiJK8^HBr? M`* OltHr\3Δ%3 KYx-* &gyt8َy'SQ%2qP(1bB{zPB)IDЁ06l` J|ih jFE$|YE 8C޵#".6'He 4/;˒[߷(ɒc]"Sde*XXL 93U ~Ǔ:, )&aeT[;N]!B%󨬯),X#fެ\?Z|m[÷ 1a M.jX[U7 &NcRMR:3H w{;vxJZ'Ry904!rc0o)5笈Rh) &sUbRr)i2mJѓh>˞?$.2њɢ.{ge ]RMϰ=}v b`pQ6vZZo6ϟý^ߡ͚$]5EحiAn]wzx=[Bn](ڸz:rkW ]^/qĠkDxRp:×B5AGa-0CQ%+0p.=71c 2hTJV`%#:N"dUplA`N# Z:RXٮY/@nc^wjaqdڴ0|Fl)RÁT\ןzJf{ܜc28A(KJ@,73YSFa#P ylRʽdO( \'4r^XoXҊD)8Z[766P]Jؕ5}Di{(׵-7xuy:R!mIOj si2 Q S^9FdHH_G~(QZ2RH琭)U&jŬMyݓ'{$">,qdz,6z|ge 2IOqy#jvHtDը!<ӼMUkmozN&2]9uq쬣g_UIRN&IJמLi V2sIRtIR^aBÑx" sD3eCI%Zj9blfH)}jt17v9qVNcu"->u55* kOv5ƚike}~ttU.~b!ZhXuKki 9gt>tYs5-6m,%YdJP*;T}%SSC%;{Pa4CB8zV0L;I9E"Oq.dKwd{E阜&z?\?~MT\`Lz2o94ܟQV`S O~R!ؔ̄*sMR' 2-{b9kO>:D8"N:`x`IĴ&m&\&RU1- KtT!DZ ؘԄ7`RfDz}MtrBrJw 20Tg }#}B8~%w4<+ݭ)ZZlF^R |Z3f~o'rd(2T-XǤ)su%0)Yv cQbD_ s1~Dj1q7ߴM%+n7||Ci+݃6F+A)+.!"N6^ h΄WɪA]='8F092=4"sS"B#3btU GHTͷP|Sޯy C=]@*whdR[Q)8獱9kщet)Rc)CLS*Zo/L;iƠ5&Of`լX+o;~LF yʜ'J"*RbW5zPi  , \q9 \i>v*R^#\YBuEKv:q>E\~2q>EZyUR^#\Y K]xI?Vw#کܴ%q~X"S;!D_{LW=c~xz&~F578=oқ\}ިo 覛E:L6콱D?5( 鱜cZ=殾i[l4<@ls/_vOfDw%4v P753ݵh*OV P9$]io#9+1eAPX,b{ ӻ_v%KjIvL\J% ]n% fF<"US9Ld W9~}bi9~2r8ڃccꩢUЋ+K,s;pU \1)QW(kJ+8wB)%-\AJRE/.W(-#*KI7WB ,U׈K+VwpJ*%UX\N*+/&vUg f) ER@ \ \K,WYJن"\p* k)8YZ}*KiZg-•yJ ,%2K+Vгϻ4-\A2\3sy9veQp5ppRJ)[zpegUsYh ƽ(!o`K݋?O~(=oTa7 wK'G?]-pD*ι526*)x&\j:gc%JE5 {(BkHꝫʠRy0a XSoOS3'ʑOJyf &Z ,W(!R*KK*KL Wo Yɯ HY駫Δ ~nxiѠW {+2I_A_v-a5 1}F1 hQ}/X?_ʣ4uwjOݨ|B8j8l;1`TkN@{C5cO *IWNB8UD{&5XKh|ab4Tߵ  I||QJJ>LΑ87"Qs9ꩲ|r쁷D1.VN!^`$hpəYFZ}i#s޵[r$Ypkg) J9(M 0sNHB(^Aib‘MQ;$Y3eEET pL4xC%IeX Y1qvڮ3upIQK.& Hi E6$ǨJ0M.% 3ɂe2&P^πz>3@lGHlp#τ#U|,8 H%ZUjIZ?iaG>珋6԰wTw4{ QL[-p=cnH !7N -RlF.J"1ᑰQÍIuvt"~Ɋ u| Vݰg#a' \w΀w58[q2&Ju+4D:0JEټBh5ֶ~|!_19La33ۅr5S\hBUlT2@%`Kr&e tf ݚG9X_:iAցtEo,z?kTujdh*˱> lu}Yv6wZ.m<&yBB{]a]H6ARNCغ 7p>~חw9IA|Z~H ^bSUi`oOf>÷TDT\ۥ0OFv2]?  qJ"+tm唐":xPZowOaVHګQ|WLmeI [I͓cˎNŎ4].%UZxfC]ת"-Xq]BZ+ ֠v VF?zŤ&d21-!q~}/_OwrFj(tPW*JOl*[:?}کׅ$cLP.Hy$+P0Lg(aæ3BӋ;6ldß6l7'KB3k[@Fb%ь'-:$l&hkŊ"6< | ٳC69:J88Mm& |z4bg;bA?,g`Kiǡm iQewd]R=; yoUg5Ÿ2$z>&(&zt@!E{+ #b~`]HhZk~Gnǁ W6fE7[Ho 83MLZA@(T!Ip>`" M Qu'JI ( X8ZDI3\^0"g;"~x'bŴP\qhN"x$D ɤ{xMn3G~RA@|O pUA> &7p ~;>G lij Ǽ~?'-U+tdSZ2E8Jp-e;qއ{t̟SV<% 碭W: 5]uN\O0?=&3A 3{/ӏ?ӝigW>kh/t ܏)/ϝv?yDO.Ղ}Am|܏fGqyԙ˟sR^O)O _rm+LpSY$U$dmJ/IS}` ظNbajSMKlz!)))vH(vNW7|}W7Ğs{WϑoƟ4pvjWu$i EG!r(:eXYh0GJp i c_Qd_Y0}f}]a.+Y>uʘu{ \/¡3s:G# bT RZ1 53H`;,WǸ% -ax] $t>5DMGZI!G"j<&-YJzZQF΢Rb@k(F@*s &KLZiKys>DI+"a4wMT~  d]/4zf1>D3-;JNK? rS \|]d^]LaD=c~2L_+u8"qv^=+s"_}6M (dq`9ozr% yu]"Lo9 WuZB[aK\DN qtE^/@82ENgq1{6¶V#],EqށWv.gEV\%d23?do!=`7Ͼ p“zߕ7N7@3g>~1bv?~oz ~ uVfǘȪi)U͡/y,u Z)547@k{d`͡ %ݸRT9m|gtL@Q΍^tV<ݽi}nma]qkUV%1~=䭞x<&e PMX_ҵ:or ƻye[%0~  ^#tYSqMOF*4AaGЅ%qgխO%O뢘3qw/x2_uSLgO߽z \Xqp lֺazj<5Ղ1/MyB)o-ҏϓ/Os;rɗyYޟtqMЊIJ:RU=(۰tGqEBLr ! dHWms7(1$)qJ掑E"4i^ЏFGS[WE(m84SI7! .HrxPm2㍔+ߔp(Kκֱӈcgm;%\uxnl3}ޝ6=O_[%qM^:5-a"%+E?/-$) \ѿw>&E aq&JTɖ>~wA-/ ?UfaB8^'=dꓹfPEfx|CI<%)LWWlb9 vոf7ڳc+ݖ!b3֙Ŕ ߙ#|8^r_mN90MNP(>D/>Ah~e@>]"!".p'C(G`Ԕ"7& ZEYC+)Tj4vCP q`9-RLp7ܰU1pV[EjcY{ O9D{luB,<pT1ΚS0_")D">a1EC(1^Irh(.#UBGgp0$ G sʍfk /'3u.uXEL*%dG(Mci 8ʍ vON#p#i甲6ao#!P8G8,2)H!c l1=`$yzS|ל]7kW^r.`9؄Q0:k;auÃ&gֱ-9|M!^JlAS cJ#68R$#Z{1H=X@UEE\mGn?9s lzD}<%PDe_"t~{/ro>|IO}QVQt*S5_jD^Õz@ gF-BZ"CTR=> ]zaހ@6>v|HX LWO|EC~휲!xlS"i!Pu$d b&08F$8Rry!(Pʍ"VFudXDEqGkZ ՔƦC4jcP OrHaJ6[܍:x}W"6,}XBM1HĤsָYPN:,Mg: 5aB9|:(ѱ."p,brLr޵bhv%g꽐3Sy2NG 3 #"Q Fǽ 2Q5A?MMq,zS+ʽ3ø)aa91pV˙]/N0>?n})ȽO}7ڄUf&BL+>K?iZӒ%UO3`?i_ RwO$iȓGWfna)pp^npYxjh2 ہ+=/V4ѭTx}^ D4Kf1.%<+pd褔=8PxiB@LA" ^-npE.ک fH{mN_Kůn|Iumor ꍪf6f<棒_^ge:$̒9)׃\MA$G(Ә]')an*oW+x[ϐJHjac!hj.٭js(EvSm}FTNqYUQ~a8TNѻ00F$ ¨ jV[ -ji}sաٿV)> K]X(AK5}#f3쟌[FL],} U 8 <; ;a L0NEgA&9eȐL`ƤG^ׅ׎ m/,4ku8i\ssnwRW J)(.ƧljA"R 1dtyMR2=ͻ"}%qd7>cUtRAa\Y&-~7?0ĠptWCa@jtrNiàFJqD'p>rbUr>E%W :U= rc\%p):rռf8GpU^2!J ,)3_e/(Qpf%d_9 4ѡհ2#\ JRp5 .zz-p5gWpdWFJ >(~·'/m\yi7U V|1O.nnZ1gͿ +I6'ޒkO>}/o6!"ʫYT{SNϷ޲Zey˧>aHv >vЮYaJj+]q(Aw&;JȦv❭V[SLFD@OB߃IB*v>xZ@fL nWc*_}w hbfpOެB&.ޭ`>ICq\\߹r<;pT]Jbj M+sHĔ-"RJ(VI2q'#4sv#o4Lu.^6uH:yɾh'==!/\x[qIE}0&Pp!BjSIm K3b d. N}MC~ [8M{ΑO:Wv۷e'b/]Yw O^?iT)x)FqtIJoʩ \/ /eX"̎}.?|,+Ƽ䗿7Ws>Gb$zkڧNG?N?pj G_vo|/_TJYA#bܳ dykav}_~r{7'pŻږ7[챽m-h/N*򁕏{8 lk2CCq⪤$}T{77_sK"lj =JQmK"Mwoo6&9\#ګ'0_ UonX60QW~3rt=[5&k&0&6%]ec5 r 5$ ~+MβB{heKnu7VKMY;+Km`5,SQk%nT8Islp&w nbLh- ELkguoC+>%Oudx9(Ւ/|L0ИRhr3ky{ ˄؝V c㋺kYԒ&kDk/0aG[83%r[-ph3ƖjgGꅜ7 0.;[JTa{PG E6L4*3p9Gs j0H=DM"@/0aǨC0 կ,kdNǝv^JOggo3 PL3'l=To|:x% Vn!eQ!ںgqKi, x$bTAof}tM(4R}e$EcHso v%,b| [ T K`!S b 03 <,:N6v(3j vW[Kq ]41 S Ä1?%HFFc3TY66" ֑|6kbUMU2Y1`Hq 92rܢP5p#;aVc'úyy[@By7-PkDix~nc(ш4uJHN~쬇b̀(x{F8Y !LH21Âv?xag>ֻ?=+>VE#9k$,ܦKj!0mkoP\ZXtd0(Xd@ӱ&@>m~Vt1' `k\z nkҩ#ldP4@ii( ̆VM/ jomda@ 4~^Y* 3ca(VHm0:XẽO70ZOWzzcn$?Yz z` qu =8*lpP {LSkmVEOp& l &_|ڊeIL-Zz٦mi\7MQl!h ,&ORrp|ApiPFt9AYۥa\66fr FFx!mw]bA0 4&T|SWvE NFj);@5?JJ!G|!ڔ2M1gȠcOҿ%U:\pI4##iH)tVrX K G)icHV´wӟ]Y׍iO ^{Zx:A=~Tmm`:f `"V% )>Z 4͠]ހS=}w0ȭ=#yۣ?piw07O so\vA޾ 'T?Is'_/9綇whoQ[\F9ȹF5rsk\#9ȹF5rsk\#9ȹF5rsk\#9ȹF5rsk\#9ȹF5rsk\#9ȹF5rs7˹fBuBb'ùfROs]e9JgSI5SΦ䳕C_X MjMdI (u2!zKȧ%+`I{ݣv- $l=]{Uw==&юWXGݼ?`?o-Ϯ=Gi|C웚_4*-!=uE.>Oy]2NEK=Ӑ q nǐw|f!qń9#koa[<:P4;37Tx%fGT@T@T@T@T@T@T@T@T@T@T@T@T@T@T@T@T@T@T@T@T@T@Tx*It tT:dT:ZŎ^U:ޢJ sb\~ *} Ȝ֑ئ-rMQU'OoYoۄ]۬1_~5jq_rr4uI(4 ϚQб4.BkWW\ٿ7⬯`!EKbudjHҐ؜SE%1JD$jLjLjLjLjLjLjLjLjLjLjLjLjLjLjLjLjLjLjLjLjLjLjLҘ_Z~oi}?a)5]cnw/}t^wk0OS=I QjNF5%1JP-Jb@RjIb,bR ?nY>Su҄P]Tr2XM1@w Q/-`l 4=|ѓ$"|gt4eW䲹l4[$UL6WAZd 6XI2kTmTdW3g?S H[oi獓)l77wW)E2 R'Qط9"g }6s|WiJYI'jCdrxь~KVEAOA@~i_ ) 1 +B#Kό92Kɤ, I,%Da?ɪ ϯ"f~>?JWO>\ 'jT}XaWVꥁ̔_^N몔Fn6LA\y46ghx"h,c!BLoۡa‡nH6O&{TVs Ŷ|+7l'k;Ջ> מ1jޖ|FX 8PD'gAX>:mcuϴwӟ]mVYYܦ*;ow ꧶9fT+mB0EHg^Ňi' \iy2>)ʨtO Q9۠tLPY?*/Ǡw#nUeUH<`o}76=?C6\R!ؔ GP8sˆT2nL^͒Y YO7&S9KFEz*ueޟ.hʞW=d,glbhT T'i݃|#RE<8+Xr/c(j~~b -CG P>jgv}@hXFVK5xY]]24ݸIEqH5UhCȜe#C .h&kpCcU>&xtw4sJ/X9\/QFiKwMaV+hqmY#Wang&^/iFK 0+h_>T >.w-i0Bz E`rדRߴ^QC.YA'ڛ&7Mݸ/\H/wNic-xwoZ~Y`rC#;X_kVkv²W){ > /\~XY|RoX_C!xnHk%{%[D͜>b K_7D³Ho5淟ΠEzE#d?8Fa4X^~ɋ}ӳ+7u:+_vfeSpe~x6=['l%z]g_~mӳ ăٗRm:)9ҁ.~v@31aU;̍n%2wCwᒌruT\S-"?VRޯm[mRo[+&nzh4,gGC^zN.I\*(v?7Ե!P3t.]/aQЮ;4Nyb'O7؜w ٤0؟-ޮçibm%A/;h!}aŃN͚bDy:fι;9n&QirvH/Qh5Ed)D슸}Mp9:Y~}$Jx}? ֯ŷ65/ŷx4|R2O[jkO,BZ@8C"'@v}6.$l {ݢzEֽI2O`G>j{EkۣHt-6F0Ni8u^*Mgͮ:SU<<: 1{O+I Ǿ;C[ CF.tt= ry9^d<?XICG#$ÝJpNaIa\9ރQ|.a'+nҒļs˜,}4DcYHomEZ"*@n1q0`۶~T\4IݧEZ˭SÏ4͠]9v[}w0ȭpt3ܲ^lTxg= cC!Gow헀m9AQ-] 4KwRTNE}HӶt4A"Or2dX3Zgb(oI==uj,ukR_3ȼ19%KI04,x5(`.IJj!F֊3C[+Ѻ/T; ާһv_ ڝnL9>bĴ༫C#>Og/s}\M4S"M|HF'e21ęτ3̣+xԈ=δG3-cd&#Oָ`3R3fh27BsPQsLlMΖӜL QxLjWs>J'UVq2άf!:_ m$XDm .@ 4T̓pPT6YAJx4r.Q.ZYoq@9V"ĨHI2ovG1A:z=Ҿgt`yjAz+OxW ]~TqP(*0d>0Vy2h+uv $ic%ߏaB(u Yn?_ bkkچ7 D(#sţMB-A,1.jR ;FjeTCH*{s!(cy]T>$JNj)JQ@,(>0̕cf5x/\RcjW0I62T*0rTl0`wL.ҝh y2hػ&{Р=y-bbP;<1Z.<]'$UӲozI++G.2U֨9&kcऎAIeKPcR^.|t=K4/gՒ:>}T R* 9Qn!ƭjQ!bó`CK+Ni ѭ ލ1MkHksߚwuEl 1|N̹i3]킡IF{U=iXٓT{:_ Ƹ,.XBR6qN>]&z|g?Y78+W l}AuXur3)+Z>5t=̽byoyܑc(QMkMcAS܋SZo_y?|}+uD30XW$m',@vY]zjs ߢ_~5Z3{^~{5HoXn^.է?zlAWk\N|ⴉ;yjzNUZ<5:nF-mhtV6ZX>yc>Y,1` ښWHǛtw=sIYtk}i^0=S|oeZzyK^M.WVڊPI|EBJ&5۫1^Mv9˖՟7ShKc0YU$܁sO5s> gnhoG&v}#.y ya rMT{qA:g9!Z4Y$<&-+}Y&a+F]*,)zHCsNKzy=PgFc!EX4HyUASSz5 M 5}YX6IziQJ(|DCm+CcILi3=b Ԃ5*qS-{U˳<&=-gڍL ي1ZBFDP2G$IK('E,Ec{c#l: x|wqKmkn\n ́Oc sss4o&Kޟ_rÒd)ї:VO~:ULTO<0BR!5$冋⻪E l4EŒu4R?iRN3z E.~"폫K7ҥ٘nd벏R糳}dq4'Ҵ_4oNBS&Nzѕ+KhBORg&S( ҍ<Œ=w݋0F% BӼiWj Br%cUAZeb DZH"DJYJd.r>'5Oe\kl}2{U}Aw+_Rk׍b z쭭^||blX^z|5yUQQY6.}ËI&)mMZ%vQ]oɎ sV[]vEUS{}4Fu6Y7H5+YMWͮ;-_Wwj_pM{Dcfy;9MpxD .@{Q/gUZd}znq]1OIж״OCAȭOIzs\&Q @78%~ ϗ%ڙ==y.3MEJ2mt<8ԟ~#t|95Fvf߉Jص= '5eoK/`$_Y ;n37bW GP|҃ȀI죲e:z-|IOO3 ;][ieT]yTemt5MK~;<f^ԬwQw&GA{q/T={Z9j] /GsR/F/oc6a}Mʟ Ǔg. %alfs4G-X%/'̂V%\"M&SK$.IhG)Ȓ=MAXˠINAL؎YA:N爙9RYbFf1FmkuWjv@:2RD;bJ1浬XYi#AE. !f%(А!lԤ8bf|H1F֖akl5/pGE0Dlm}S"kW"N$b'W•[ @;?zTb`zt{ɌFqMJBVhT~nrVa%.kE2r3!h%j1i7@mgh)ҫ$dz+|dlmW."b'׭O 4m1%(_B8%"J L ƀ:$K0!E:+ey ZsC?:ǧ~T&?޼95iY^Ъ=/N\|.R/'B`R~H݄r+r{y_\嶹vۏ8y}'>.LrʄɷU:7ԿOSO2}WFw68 SmpU$%1)v|ޱNOпUlpNFyZQIBU_+gБ:]"ʭ`;y :Y#XanAKMTq!KP9L a7^ʋO/ۈYIFf%d⨸&lEf ;.`B;íh"0z+}0嬍RQSQBĄcVw=paXWxG/V+jnh7,o;|DX wt:]oGWY#}0|r!Nv?Bcj 9HJ"u$șfwMOWկ؃=ÔW o=mP@"2G33T BxqAsN;wOoڦ>% xer-ZF %%bv *A?#9m]*|eg_M?C2=lmd2QŨcqmT2cl = x$ E".A0 C 9\8l1˃KHJIT"3 "YGZ/G?HlAG>C$qN0wZv`ONQ>̩a\7H2 ^ՂzR\Y%+JWNEnӏLGq=[E y4ɣDI&E2zes.}QQrTrwf!}+A"y2\waz 6 g,9>{GzL:-!W1Vp7Z,rPwwz e^zO&Br涋*DA4EUBd< kx#\rbpc>H"g\u=.9銎QP>-$R4xX~Rⓛu}rDҨ?(A1Kq^,!o6r)gkr:Qbmbf["rgBƖ>J"pCp60[2\~fOHxkm[+@d ^ǣ:ifK,:}7S|9ayc۵dۗ5~7%W?u}>-~?N'IǫohV׳cw<ϟ]fbko2=|qT9D~7rZѻm'_ON ~0oDd~Z2ŋ4G-P.p -g/5cVo.1ur>9yU]XNR*PHY4u/]GR^<[u 5RjfGKvC[-g;9EMŝ*O,gP:n[v=ZvսcjvEgߓv k3Am{&ו2Amax=m{:|,u؝^2PhY.D?'497.W<9y؄Ck DX .%% bdMFҧJ23,PR+6@ӏ;~cqGˋ;i@QEpBP킠6A%Dq(PE/S!W`H 拓])qڸ0a#.yYף=g(!򭍔[by2vZVJPV'wCP8$RpaKTTDGd663`j)VL-MWy.B2ZQk"A^L`22픒2 =`YYeYg V7H-FltPM9.٨8򭖑j0Y$&J=jD۠6Eu_3E^Y^;l.v}$2bLkfudTӪ4ՒҺJT* a4!ט֑:J+ɝv:-[zcsm+4B iJ('qk-U|'b.H/l\|CxQ˺)K{|  &^CZ+>mW:UbδZH,ϧ1J D+Fx"8pqs ™v'pTDEd` 6q4̘8%3BSQ1FMdMJh'ڦٌtS>RcK Xg\Fg}N1 d$zS\Eur |toyx߾<1r gE?4TMiu2@>ҳfZ;3(Uxc̔gH\UENWLVR$aK##)K'#r!-GH;? |,'x>Y037-fݭжSyv&B11paHM|Dl 4>I"J$Nm<nԚ:<>A˹IQ-Mђ(zF۽s͘:нrfJFDZ;[.r4yo>ߺ.hk9{:kՒ m"L Oh脉<2QŨ- sj$<_[xasNv&V&W10i-y H't҆QhoMF,Wq7zAeU(W}*.K=^2ڍRU ^Z Jb&\ TI?;°J5]ӄ@bM>jt6ՎD|+&H=3mr]>z/Q UkAp&URIe742)(gc(&~ ~SkBŞ": /FظZf֪¼bSH0$7MQfى 샙Jk]ptwjSBj5I6CqΘ,B\أ"ka{'dP+){n6c?>ݛQؖQBPhk@Jx2 3:+A5k-Ra9;N\.! >՞g8qjpfwjI򎷗ʜ"$C9µZ?e\0|)@]q*J1?gm]4ͳ;o/«upr$1xtx-]sR[dA^:rrrP=lI# ymðhurjR˨IS2kI5>&zst<_pFlGrۨ۞KF x$,qc2 \?sFh1i<.1qrI /Mg].ǿ7^}śW/_pP1m#A;ӝ0\_}wkhqи|v¯1.}>q7^ ڠnw?^ThriWޠɛN+ArU,bmm-vqpMPJ- DZ!1@O5.Hg]]JXa( C4A)(i.Йc 4$r RhY#O8 8t2翻uBJR` 0>VI`\DߧHpiQͨ|6vz٩J[P:yJnKk%|C jaz&*ôt϶\)zosRK!++`"3MVCG5}Bv<\ء@uSC$T#v4DCX[L`Aqÿ|'4QA;H yDӖk녥!R"1DҀJhSz}s5̫渠!V`7\\ op'Q;[R\*!u sB.#a1K&_EKbrNg.-W(gۀPϻ$*qRYKMr!<\c$DQ 0b")wPK=><9(x_&j)JĎIL89ٰЃҍSa{}9iy[>>\z@sJ>K,hP୍c!7~t3Mo#XX#xS,>AU̱]\N^Q5Nr-AW߼_S%4 Ӿ)@Ny'Po{Wxky~+XsqÕ\PkNX&hSZ:tO>nO?ȓy茫4@}&ӘVԋ!IO؃[7{H8# ~G*HyQ+RAQN9EM")ˣ (:k>r5x߀'O[`jH~ E< &m`6*_3iyipҼ8jJs7_?u\tq8U݋'m|1fW߿+'[ =[wlX_y^481?S)n$U4R)ԥTIH"4TEw.nJPܔ⦶Tda2X^sHNh8nb24'BwpB90n :(BT.z0b T{)pJ;/rO2t e_F V_gSUikU̪ooQ t^ڐ%Ah%Țrə]bjdm-~8sy<3!R>DJSB8BT/$!޵2EĻ>{AM5IJ?EIP`._TU(bQAF:ML8i,Ikƙ2"!G B;4aI0!\yϒA AX[ֆ%XNl6@xPB$2WY4BQ"D 1 ֥$^{L4g W/A~bge;ZFfZ:Q>RcAFDZA JpGVkPOIR.SC#HN& "Q&xBEIH`&bGkU'8ZU5x!CeQ[QSchBv⎛)1'8+VoǮWui@"ظ'L+%%xЏT0rStB$/)Où:m!2E(@ZZF*RΗN:S&&97xk f:球Ehi 3Z) * E;D)QvwaU <0a-ŽlW:=J9؈)+|֧G9u0Cɂhm9b A'zR֡*#s9EVB+h e(J8/̊RQT@ldu(o?)v5Ӈaӊ)z7y|MS9OcUF+j)ST3JIѪYJ(h!j!Sp!ZDpT(1,ZB 6ӲlYh/zc.y/t\c5릲Zy~``~o )ޢ+qArQł&XTSSVUZE`XB;O&QHLr"qC:&h`$b"МId$ΔDbjd dĭ ~& ١էܝ3w,a;=^K՝w(fo1`o,ٮOdjj 5Ŭ:qo8L4.mWk5hi7dhZJgCwۜ0wK&Nf.z%oZ0&4I2 $B"T&/sH)$$m!mq))\Zύ$)Nژ$J"TOqцt14m4pGs8#[;/Hf r}̸#ְཞj%A'>0;|/ƦbGJ J-/gT0AA@zl IpON:n"agc9LPM\"r!(ΥՎK ehpuXg#[t7sw-_<6?sNo``dz}{#(dQ˘T2"!p45h Z{Pe'|W%2I|R'JC\8&T>ʈ\#GZ$j J#g[2deف-6w֌[X# pDi#aY +74QQJms\RvVrlY v`49 Y+z:$% 8a9.i@T,ZᬍB3FIrYOoU>DqĒs=Z,hփ$tZFrd~ &F{41pFD;rA ׷ zk&Xó6CC @Qۖ]?@l+^jCgn;8ybpJ" @1i K)t #b|mb][-e|[r ?Z9DE1FpXՓ?k53?,txmBLeY Uba. A篗x#\NJAUu􌭆?WtOEoP8kL>佊~~I/uͳf ͤlH͞OǢ7#Ɠpm,W_[f9V8zæ?&Ji~_XlMXFؼ'  רņe+\p[B.Nd0B/qcwZ9-~WBŬpGܙ7;CC;f#9EMx2[L/u07?bFKxOSɅg\:22f>d\̕X CiSB<)C[Pڌf8Ҍ.)?wGN>*MyNi'n˼,Vp4ȱ1l)k<7ќ/,VD 1mq\+ րNk\ܙȫW+xUQ<02%*D1 99:&WUNK|7o[LW8>*:#=uL]M5BZO?,.~G` / i8oTu! TA'#N~ ޔ-IFVqV>&ޖ ~(]O=!ޖH+R T<..*JH;Ȟ| } w~N+/9TB"hzPX5%4$N2yCHPd %H@LD$g) mq c\\X!VH>%9im8 \ g ]9J4{ YoJjxl7JꏪaFiÍr)7£21/|)hQOc5MB͠UWwm%Q ޲[SxFlɲDi|ׂ%)Z&#|)0%@Jo8Qƭ$ɉThP0-T.LDy A)3'SȖB|:-: i֒'Az4P!MF!*.3\8}-?Ͱ)Ug!0VE)*Qjo"jP|J&\*/bT9`I? +Qcm[M) O ÔPI-P rmI`@\F@>z܉Zf]rsm\Ur{\ qNm4^5p&p+)Rb2ssgnHAbտ<.T)"_-eٻj9Z.Vb.5\h,70Z?>DJLN f]q˘,r8@V!9:Պs_'(/bߔ͵oP$ZqzBPEbҞh+Ap5.HϿ=i]0VPW痁ͮD[Tc>=>FU•aW_-NN,J@PNñV˅oO0.Om?\8CR=-ddB+n̩i]Lg/Y>Ÿ/'g?.A.\O*{gӵfK"$.'uяg/n?V?B=IOI_Nbc1 v'68`AIvbb;HT,;!&٬UEtV6ssM%pڨQ'6..׽ Kjuջ~E͔f:R4|w4Ru;ʃM?QJdk*54&Uv6n;oO~~ӫϷ?/:'\yyՏ4z]Jv/; bh^yjڛ5-A4o= ڕ+hw؅!n V׫?Og,{gi:-5_?8{ds$< pZٱ[1f.^0v@Rֻc-G%HlǓ J eđD9 c1@1!~@c(K #){Wf6i矇p IBYTu㯐e VSo.(1rcJLs ") ^V~:[ N]s S\I'!yف5PO']\JHw@Dв~=4sMq )|+Qr`\ Q5q4ƏRΫKݕu`b^[!{,KAM."6{Hg Rʭw=.5]^:(Qiל洂gII\妤ulBJ?QlMX57饯Nmr`m^m:a_iDT%H\TrG1LQ)O[UL9sM;r-+s\/6-(ĥ]Cwo`ڌʇ4(w9Zӻ4\R2~9!'= n?M>zwѧ Ud_k[rsɟ]Z]Z/ƬVe^VIdrHm}~>+!+,[!} vCez3w'jm=ꟿUxR<@$Q9F1ob{"8oOοȿX]6HRX9')\FP/&tGh~~ 4s?0>8n_b-[]`l廮t'^|47أ4Q8Xrf T'̘XV_'\˪ xWR3qCr=xC! skh'jQT~ ^G˼j~~}b^=SDs 1K| w5>vƟ78<..L\ %0iFS L֔pT''8|B4^8߹`YwMD"^,c(F˺hENƒ GmJԤYܠ GxmEf1:̑LXQzW<60h5q6$kMaөboP[n;9Z !]IOj)di0(Ȑx9U5)f(gqC&AȒ$ U %Br77A`-oKӸ6wxU'[zMk5<fdZ|լ;6J39s6[_SY MfC{yeKgO+j,rLL|E`F D6e>h9ℒDRZF6ǘC`ŗ, ӓl٘ @9+oS4HϹV>mA&jX-62PBo j w*xE^w]N_yr= 4|[.}Xa (dޣ }92ܗKBfБPd2iB{܎&R̊=!̦$rlT0d'QL܎Bɘ2bKug]M-v y0jVVGFuAFTIew+PҠY{^bz$1dB\F,>4" @C1ϙ16"DZV~{ؒx|S1b--&Xo{6%ʂC f:"M'r^1H-D9jxVl,* 1 dV&@$&-)S-b5q%ԁ⤙*!uV%E^..vqm2cc|t y4L81ОQp !HH$ƘhToaVCQLX ̮'额#7Jv uwёi~}iꙿ_ ?Up{B"4B(b!iLl(`_i/C=N~,e[_(5iء](() CvGCE>%51ѥ2ۜ|bBEGHD`NtK!E"hbp!-#>1Yh{ Pq,p w/4BF\#zMIK&$M0`Eg4d3ۍlNAN;Yh%0Q21YiK(S.M4I+C4QުxăS*Dp%=2$x(12Lғl$|P;8y qJho [TNW&f!{sGf^{ lwrà#@:gtH]LC~h t|&"WfIȺ)ehw0RMY|(TH}r܊( _!wŰ{ir(||pjoLחmiq+s"N C8eNQ(bM|T&3z8էSh ZB(.uK8Ȋ:XpܪlQ #I A3==ؿzu8fz: o^u=ׅkE}jj>W]j}g$\0UW*PK+zJe?wmoRɆJBe6v]~Q+ATpo\?JxMF9n{*/NoМG-vo(-w>n<)$Zٓҙ'/ң 8"t%偣1!Q+U8KM_&{Ljv777R>ERNRStIzJf$J+ ~g-%|_-=zё̕.[tPW^H}toNf˫vktO]'ϻ ;[tچ[׋Rq<*1'50P.܈'}A ^eЁ/o4dT `l@qp=E0}$<L7e"\sc9{,Hkա+)ޛ\YəwhB+A<}ׇG㖂PON&G݋/6@F7(s5-~w]7f[&H9.]>z$1̹7A thl6J){Ƒ/I;R_!Q' lv8>pb}S"Jq翟EDJ5tt]R* ~ _ޡ6u-K:6.'_6ń^|Qt_ɇ9}}닛{ 3ٻ.WMn酛/K  q,Y5{͚]ٷk= ] b I"S_wae7b)8 _8_Up(&P;*M34`: uE#ÃQW\nE]jv]]*3TWVp aolKfׁ|yꦾr==NR4  5=%O\.Ɨ[#!-7A;bs&8׿5ozlm3Qְ{5q.K (^Nm+G ҾeLFr䓙T+A@|˻lP\nf13)BU&{,0ݯ]uXd:$j,(p*^j {r(橡v~Z(Q;PPxץ3RW` * %\v]]*3TWBI}uU6 E]j-ﺺ"*I/+CBWDd r9*ԊΣB佺z ƃr "ؖ*NPkꊨ+Z< ueF]r*juuPi{u Օ֜ky@lG]*ΣBнzh H]J؃QW\iE]j躺*T>vjŭ9 uE F]5*Ԫ 6TsTWV2\Zjͳi%P-ݔ[0U- GBc@}hy[i*+pa\ qJsj@RhLTN$pyd5uբm6oɎۖX-`7?,tq|c]fwYl ?ۻ&k}/锗7q&ۀGvg}[}X&"~X5hrOGɶhviG-HGʂX Q )pvzhkdԘbbFng8>WQ݉)E9%OIr䚄6H9%pA>QiYV;JR Ke \T,p-",7Wm̢厣pQ ݥjǾijxO/I]bO0/<=BzD [Nj b6n]]|.4LAX+_ڇkftRA=L8vgL!i;3- 3(El&5T20cAE- l1g+XNL ߔ *'#W̽*:#8PځV e9̙^-fKiu z{lUP&Fsէ«K}>!ۧ?n9|3S!n_!5*([-&T~nysu7JL3.N״fO ; R 0A  ?_zM+W*'nus%RףNhHXm|&rUv֎ӭ;z*$硚xa* ;~?^}yÏ/_}Ƿ߽| |{W?~GV+z)*u$(7 ?E=3ֶ547Z"<gg*12$T ln qGjf]43!۞mt2KvJc=X"g9:=CJDT.E[}0蛻W |S(."W,*ZQZDrJB:gF",;H%݅ ]|bRFZI:)dH>,Qی#5L;xЖG)\n z;TUTBbNrk9fJL X .1ȩhsي]Vj!T,W}*O7eM8-gFƓnM! za+x'&H  vBKGE XiDQzK7 @[(`P|p ZOa+-i^r[A_rv, Tڇy AD+R'V9KÏdbjpQU5FoouK?]ףUh_uZ[v+7GFwݴN;-^ ۤ٫ܽrDc9QA]Cf}+ႥU֠[}nc C'z[~C3ޓ,E!F]w'%rk9XWǦ9(Z39Ǝ9V˯Wc' g2^ڔ"JJgc m(^"Lhy,DdL+O.ʨ5H$ٻ6r%W|Y ,63YLs3ޝf>mmId' }Î^([{L5fC-d1q .D8 L%ʌbVhG (ODLblv謯/^?;8r? W?vnoY ]‡^O'e1@|K[Q|֌W#>k&0/\*)jkR!ϾhSɋz7T&Κ,H1 B4[&H] EKfLz-:UE1s6{8UskG_q~y`^2a7_ Bc+6 q<,T 08T'n $d$$AJ=e2;4 4]J ?ymb^1)CRTh8cğT21 f[6D,N@mX[36SblI(BŦcjallQĖ;%Z3MG7J%=bgg2"`!LFq6cN:jD = e%͟IG C'ь|7UN5_QEnIa-d'pkk [Oh/qH+2Zoլ8;q-g2 /Nה'i^^ْYS=pTXd1}N #譅9&Y<} 1 bqKDdw1"bk%(Dp䏖x\!hb Ғ* O`"Zs3@ Qs$)ZqAD<ȤjG֩ˆX̜͈4vqC,Kn0.[\\+f 8 0sIIKLe7:8DqZ3 4}aq[h!2Qʜ7/K.Hr5X8pRml( çr:kԮml88=EGLS{Bמ$Y qg$r%NliH9wFaRGgtqw]G/[aroZC +-FÃ]h\muOlPׅ>s)t*ᙯz'6m8Jж*Ae1AtZRΓɅd ER°\PMQ0m+=?w~Yu9~l U7UKִQߟj`s \ǧ㭞u?&WmtVfk[KVXrUfj5w5 o<R-j^(?dMW?6}~+kTf }.;Kzx9}im[^tb^.W> xr7T{Zn~O\%Mqn޿-xsgq9nϫJ RB7`WP e(*8̊JQb`ԵCWr{rBSTb "02Z|K)$; ܃g20m=@ v8@fp֭YT_z7Kr{|tޡ$.HB *d4zrAq[-ZE`X2% `u<ǕAH!'F`c(F3ę' PD=g j=+ijrՀ8s~PhK]j*4:j}tV0뜥f񖖻w oىF=vA =Ytި [UY+ai.{65ԠlS R4!ρ7)J*KI$hR$m!1f';p))\Zύ$)]N6IP򲮐O !?L+Mx*3gs4wGZ{`5#lӣqCWra ؆>cBk/o_K$$U@T"Z^9l(a?E#C;;wgf:^fcA$!p&.9?fw*38S1aGg< o,@fw7:3pe4<yDoAXIIiǽKsCLq&z{ V>/G Y͵(ڎ%moh*&Cn2v0ao ^ec u'7<G:y-~2aStNjn]>rw3We{n*دĻY՜1Em"-˿qp/e?k̨˯5ybBcj3qsjDI9 d SM1"( %&wq }dփ!1{u;f?[O'r>Xi=u_>ǻ1M'b2f" 8H<%%·[D "|S0uHI:d!-m2b@+d0` DŽ G+/2s7DWz5Hޟn[^[Xb*j-'`A5Ddf`jDD)iaYIBdC *"#"YV9A8 JjB@ K)tiZq9uϱWs ,0LkBŸι) Aӫ/:_30v8)b^K Y>vl'ϻOsˋclufׯD=c{7=pÜNwP=6}%i^Eҙ|~Io!Cڜ@ٝ~NNw\'X:x]sevLO9E6ǷIjmoJmmHp_pzyn\fñ5eZkmJt8 !:x:WX]tO&I %To&)7;淪Ǜ_f?1ݰ?<Ǿe9g~?\Ez_êʿmL^ſt@,[i,,?yh^RK>GdAxPQY=? ݫ MjЧ+h+-ga651|4&|&˖*~Y.Fin;DxDmJV;@8A@z::d SV1o<*T8;OǏ,^Qŷ xޙϷܗ,D+ j<DBEF%ģ46 I5跒yqmœl=ae]N0@E7`\xXuԬQ'#4F<,CdsN=pylT.Tx\#?&B%GWh"X*[ɡZn•\OtG#*[)GWݒ] w`p)ʣ!Zud0[j>JR)?&2 p4p5X e*[I[zpѓ4<5^Dg4/?zcSc:Y2k!WRR#OBe,U ޒi zXw&W/?sœgy{`ǧU9M;oY8+EZVFRj+]mڛ_^l]Γ]^,pǥQX:eW~Ӽ66wI/LQTJxb*'gL{7O0-K2ì x*_lKV^L"5r 0Ҧ]𾓍VʊK [㚕̝V BHQWN +\6"yW.V%b׆$s|uLZ=ލpNYm}s˷r1?1g2y0+צ:T\cK]׻c Gǎ3c/WS4!Q9OpJe}ϸ:U*BYɸ_]Ur_άJbe]^BY?C_>+d=;"r]Ԯ֓~r.:'O7U6%Ee&W5q(O?$8oǷ1N" g `^ aG'zrNv&:i@asxY\I9ןYÉbڏlXLVL%o} Oey%RFc(({7(_)9{Z~-BK@.Y*8)RHscX&KD&=]KshyJB"]JչǼ'n ٓסOzyVtHwU~hgsQΝчoWKrgx_ g͸ٓqj8|^eng-=Krz89 &$S'sd\3'!aU 9t+P)Zhh]\?t:I(E,W] ".hun'ۣV4i0P5,P(^zPPhC u8h% RR媓r}XURwlrۣs~ݚjK7yO^Uq,s8 >K!}"Ie7l+~R@irV)e)d>ߐarayePoZ3ߴ"RZ".5dM&mr fyTd:)b:H^jM>L9edky5$;+l߮"VOeN+9ξ78;֛iYZ":˶ؽuq;Zf=7n}e3\c۵Mhyu.Zᆔm`AXaUOPD$WnpE Qa+f$ B`n+&BZ\!Q:(WB͋ m8AP i](W+Ij0"w^OSzN+44UJ& He0r"WH{sN+li;r 7$Ws`2H`reA+6$#WH\!%(WTk+ĥڑB(Wݔ+Q8]%56C;(pX2, lSB "PYn%"mvv;hos14C9~V6Ԅ츍W`[@vmj/Y*EϽmR-KWFq-7=&vK%JGm^6kX O˯\C"Ch_T0wi#*̵f[mZ>zm[VQ67e )Zw9*2Vщdj;^yYGXkwSjYTٱ:pS|ѲC)UdX6noS2CJ.csb+,B\hZ )rAܲHت`SI^taIrT.WHhʕJP % FWP irG\Ij6+N q EVx/WH)YʕbTpܕ5gZ 3-(Wq"BrWLY8r"BZ|+q0E2LR+2B\C+UwBJ'2tQ,SZ$WlF#WH+mӍ(W/#Wʻ(1}jzI{G=Œ5o<NnpY͚MiHiRi@yT飫^^Y_9z2#5^s2_:jVVB$f9G,A5rAlY0pŋqOHK)qƘ HX` qEz?uqE_HHr\#W+d(rR.WH\ .ŋN8UX~qM0Awr]uQ$ג:F9rh]R(W+-X u8v54BZ+|+r҂rB+4B\d1R(W]+#8! Xfq E (G\Y! IX#W+5އrEP)R/iXM]1%j[Ђm3`pe͚^u>nFL(WqR2V xs@xPI5 X>'U_v܃̀!?wul:$M؈ FPp|+̉rf4~%W f{BJE\uPǩ HX` qEV+ )rARi+6&B\̫K5D.WHI`re$W FP i]P{J+b HR\9\\!-R 媃re{ypB77HFU*\\!'v&/W@ɈrEVT>rLdOY 5uݷ:J̷tgT\ȚSr>SC9ȤY3,wX]- Rװ̀eA7m{ B3Z&(}(@62Zc*iHr#W+t(rmxЌR(W+р FײP h R(W]+Q$ B`+$w.WHiʕ\j\+e+R(W+4D`i+U*Bڶx64q0ERZ۠B,qW+qWHd(\uR}6Uƒ+ĥ4BZ})}[EJLHr`P i󮐒(W+xjVL5LE\4QmkT#JCeNʕ7W3hfq*t{(kZo~~/ _?^5/s(:[ 햋:tR}=ǹ$X\TXn5,wb]l hyY?ɓ\]_ժ3A,'<)! P6 VtDzrҫSo/W0Nدg*g6zqKܕ4DE~f4*XKw3$91{Kr݊vhWOYvz|d>k從Rbj)A[+.U5pܿ3s1Ke<*é2Qp*jFFhfphwߌVB셊j0WqlS \!Z\-#wBJ\uPpm@r*B\\!QRZ媃rō%: ?]! E?Qy+a$ ʃ+m{eh3Z}+4UJZj He^2\! )\)˹> Z\ ` YS\!'vpm(rtr&UXU!B`N qM0AGk}+Q(WMB+Ld@p"~0qhGb}]ESj>=8O}ܷ:J̷,CW?>kYf4!֨EeXɞY_(!T0ݶ',VseU"q ' 5N^34T\>h C Ay 4 *7˓7Ri"CmTw᪹ 77>ox3~T?\gRLmf "Ait",5 ]~\ނOqzˋBOµ7^|y?a fžK\|ˀu~S}~u9ޙeBտo*coT'fv_.lWoJskXV umRgtڿ6ȼPJڒ/( 5̇VʢBɡ:RL &8}3pټL4K?IvdL~IYƍp1(%$4C>(`oBS"聍d Tr2=˲|XIViۂl]4Geg[&kwu?L@>XoQeIP\8:x~}wӭb}Au;RRks1wTwΨ NqFbaqamjP`#6%G~XO!m~kU`( J)yu ڠJ.v4}V~A-WQoh~q3xg徹)+C~x* 2OP$oIEˢ$ T~:X˺e~28Vʼn!>qw]bmOl[c|c<ggQrh?}z{mU ӇmtZ-9-Yeչ*XiLAk婁zZʜdPC"hvةfAC#BCuSG6Klk18- ˁaNQ` -d`C%,+8S u)XI_MBra6Ho4usm~:>uU$|}xVq`\\g}ğ{y~opK!nfޝwh]koF+D>Mw떁bXLc3I0hV,K${dɭW%[v@b,dSϽeUpF:tqYіcf܃{RK4{6#^1FޢCf̭́`\[y]R[헟/LO!<`EJĸ cAH;q̚(㙗'$e)Ѩ14A⚃vף׻1őhܻmݭOGjCmo6KqSFףOinvqU`¨Oy /\*)jkR! ]F,YTc) ,D#k% RTkƤWܢS%Ola.fn^g?=m`^uqs\)fTr3s1Kc$9W_fYŤy:,A(-qgZ5ƙf;mݟP2~j!RP*)c)y5N@sܥ$&S2*M,E?oM[>/^Nl討1n(n*K{ΈeD:i !LFq6l)Aa(Gz@2>4Hm"NLGW)~^CѲ}w |.#(=JՃ <?+xܿYx[S+G̶/mZze.er~t9.:E  !!ÁiƁ&K3B(я# 'ރ/݃0؛l h; <)h.>>8E Kdo-F Wg\MXhc*Z3F?1֌Q`y8$(x1h*KIô)D&i d|RRƹIRVIf scZi++e(! d |q13-~Bg }-Yy)onoǿRyL"u5׎3[+Jߢ w5~@f®~OډyN2glHR[nd_OjVؕ5mo߀1ꔶ0?BLr4Za@ƶ_N; jɻ~-$|=x==`C;? _k`܌̕ O}Jڹ qV%Nep,_rp/e?ksT |Fz)ШpCnqs6p:mRmg"qⱼ+4pz84Т}W?65Gk>V{~jd _|o5S 1ӛ`2s@ِv3Wa8,-gqVao7ㅩH .6N83 P~:%'{3\Nn L0(%`kL~ᦶ֑"eNpM:JG(=V) %ϩq%xM d SM1"( %&Gq( Ja~Ng!] g k!Q!凂:`x:`H>"(t 861IT20ʅcB飌H:XxMWz;H_\"n}lQ^S|x0 Q|h ,<+74QQJQuON" :)<2tгz:ymȣ5DI&E0z?0K4`T,ZᬍB3}כA[bpCA3#Njzd,HB:x-O9(򄑆&F{41pFDi鼾OwNYWC >oQ"p.((*zn0?/0PTij>?'C!*GGEdTR$*'hCImtRac).0]_y+V_f{_Kϴ t(&xw֦P?,Iу{ϫooz~[ڄM}ۛAvF|ax.u Qvh3v0z?_p3'zsB!zS]J*jd'sCڛY@ٛ_?Kn&Md:p7qWF/{{fYdhosԬ #jm6\W>mCp8[8\;L>޿|B7=|mgt8FV' Ey_$T$gf~7S}L8{cھan_8ϟ?f;>j槏As..qp:b|_g=k;dvW09C:/^aey94wqa~oG?觳Pz4&m?SwY&wٓk4fK:82gU=LB+eh#|~эsMʿ#FN.G_?tB,Zl-99YV57]gvYŢ.}vl68jƽ ֳMj+ϋox@m{GɰMy31|2&0[ r@{_wN,Md#J̏t?, )nxp[HѦdu^R8A@z:/$X`(xid)3?91>Qo9~590/XVxb-k  Z1bJ !'hgQWfhW {y橋NEcR^0H#Lp g\FWCGS0$^o9{L(xLgyO7c7w/ h ;gxkY]1)*ޠ%"L KVa8OQAP!(2znw5Ht~vz=7̗GAɬ8~Y$cmW8=H_InN"bE嗖} NEr]w׻%];s\)\ۯhҧӋ>?iaz+Q`~x3 ^ :?K&=OIlxwKGS͘f|l沱}xQɏW7>hlը[trթ¿>_7>fߏOu$}}&~/Xu×:wƷr /_W/}/^&~_ Y/Y$S"p||<@_4/{144m4i^~vpB{ :8/:UT M [.&XA]=\ʢl>>`?t!TJ%Ge!~@`9ƍ* ?OEj[3fs sG1&_<鷞p؄=I=,tuFB] !p܊ʔJ>WpR&(*YufY٫M2Ľ6\]|&}pb{Ξ\RD9ϵNrtRDN4˦3Ĝcߊt%QIKY5_~b\=/ {w-O1ϽdS( )_KPF* y4)g] ?tBuV'ŒHyyMv6])j)ZRk@YrlJ g}s;ЫHRlb9r5&ᴆ̛*RU)N0EBT#Ad<%=%dI%xC8LXf-PR."r*-=xlB>UկY{ZZ>'M}D%+H ~E[-V)Ga~Xʳ>l3bk !V%7]A+KR#c{c'u\{D~{8mmgwO=N~@>r~XAr9vvXfGg'o}l%YU3-L %~>jP9f!v!h-r QŪt71o.B҂ڗjRy}h[kߵ¸>߾'_%qeM]-\ϏRǧOCCgՏw.kn>bμW謟2뼇>ppmkP8P`jGԝR#H0^j}1c 9>)}p:ҌO?}2.G>}5^yN_7/}UO?L`_]௏?~rz~R&~IenäKGPdNYH[./u?qƓ#z_VdAGg1*yO6]d^$!6ln!ӟ9n1~;p4ͻt _.]g&p<{5.#_wrAߩ}z^WFtm9E??j~'86.c*mIεeP/C-#uZ֗IޓGe-n/NOIlVa?$XRJ)zϮ[fm$6ZTcξiרQ%:BfrxmsR88- B/D}gJ W.{,+ v=zȿ7b_\Ѹ[13s#0\<> ٽ zYe^; ־GWPMh t$nnw>蝍ӥcm8Β&Onr&Z$kw7k OA+.U~!~az m}۹؛E ]kNcfG:- p )oV {epg5ߌ CvƟ?d-bcp9HZԏْXˑUlͅd)'[^9p꿽__=^[5i}'IT=[b-V3-iQ*J8*GpLS|ˬB"]8퇒}49eҡZUbC7(xlQL)bXoUqI1BS$k[ R Z@nNi}VX *cds FDUtDߴmES)rv>7)-Tl>Ԑ ~ʖlS1+Dl3WOTj1'aZJ-&Bq g06kl,.B76i~&!Q5&9%|nh cHõ>Z&AۘR:wʒ&S`$2ca<.]EB-WMy>;7*#V؏|h]!pTɏaz*Da1fN{aQ2.LQ\O@HB>Uw\WFΔ$6ϬTE* F[5j!ȵs26指57h7žG<&jdI!Ǒ֑ ~ۄiҔԒ8fu0 B D+EbOp`R0 3@e|75d&v/:Y{S b>R@Hqps$eIaj` f,)aHl NJER!\F*^ªYQ@ (buR/*mr 9ު7!_ʔA$Dd(5s eq6uҰ&T㝀BV I@<383l&tkiZ̰KUT3f"1!D thwD1!"'Dߛc`;wcȷSt#P,}V2){@:I,*#`t8dAhVaA3¹~2~-9J# IdDN+k^-lVLtqZXy@ /!!d<ԗMd :Ynt# nCfp6⥛"Y ԏoT?`jXU2qf[!^V`$ S`š`|ޛMŻ"Q)/сwK |CPB0- w ,=k p+ѧFK +I y 9z6^vBY1#'$]o/ =IȒL\9v맅:tf 3x4!, °n*) sU2B@:!XZဟ?au!UPg&)Hh03IxƠ6T 5}魨~G`޶Ia-#!Jwke \-y+h{T_R 9<-8X<󼕛Ch9=X֟zLoߨۘ:3Il@"7669hfѓRpiEw'%s?*\$i@ΓFjA%iA ? zV[Gd>D5@%\Is G+cwGz`ޘM^O&җ" :ޥW$J0%e?m3M`v2dL3&P ft`zm%| !9s"`'kMD|gE4%JRpaaT9d Kc8eX pftG{caUatTFGkrs9:"ݤL .}"(RE’>,+Ɩ,!Qr05L;ig1,ŴvbYL;ig1,ŴvbYL;ig1,ŴvbYL;ig1,ŴvbYL;ig1,ŴvbYL;ig1,Ŵv΂1zyp2ojVYVu?> ?iU$1[Ufe&lOx@>iȓ3qយL9:~C"J[|'w7zqxVu;k^@'3G޶Iw&z:d+ xZZe>uQ36c:g*FXBhV66*M[70u TK&P:*h3j@gb,!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 t>LGZdH~]<,+g@7o@CJq2/Ǣ G@%".1* qV-%H\:҃+; `m:xuh5! JJQ] !uQW.]QW- JAP]pӥ,X8.LV3XPJQ])NuUym;Iw Z}ꪠ4 +eC3U+ꪠUAi3TWݹ:V];+ꪠ5'Ұ \); `ޝWt&U'3X4uezW6Vb~kiE TJ0d#Ua̾}X Cf-O7VAD="_GY_s<#+ xB%T0|vs۫gW7rRX}zaʎݴ?27qG_j!y7CBk={!*b>}w5M0_ )|&о*Yr?, ov) `.- hH(~+h(̳t|u 2&Wj=&otH'/lԒ'Z*ÜLm\QA-2 `m;cv$?v&aAyj1hIʎj!u+p&RW.7]QW$j+.C Ʈ;ꪠUA ngtH]\uF]3ꪠU' sTWd&RW(7gL=  ڗWWP])9RE3讨r/jvuJK)D `g] 8quh_r*(FuuH-yCU$<'UAI3TWVZYԕU5++ꪠ}mUAۂJ^˜+ N1E+pMy/{)`V-D-[XB|P"&M0=fxqT Vʾ,v2yS痪WFz!rVWzF~x0*Ƥ-M>R+RHr*PB"<+KNBκX[-ISX^(fJf#Bn(kRZ#$g~뵅g}cJ6a|DIOkYYY4zRWQWWuUвWW%gP]1.J`tUIkAoxRnր^u/.?%\urVZ7>y>Vzߘ*᭩Dr<*nax8T`係s?f$6#:%GA\?묛ԕz*(kf;m)ĒoqK3kR$TѣŘQ3l^2>APhsz8pt3gFB899zn?=z^/As3qq >} DźeM AIq zr#'REm/^(ֺ(NfoyPOq[m7oY p5N4HAe^7ûyeG 3,Ϫ\~7_|5bݰuCػP וSԼAPl:7 }S[Y9kvqtd?C*-.>ԫ_z-bsm2`fַSAY^-/lpc^3G嵫]pS.Vۢldiђ7Mp0kKs5܁9ev^_9uUlS,I/')<\ǪBr!?q JЪհԿ6ېh\(Gu0- g앐Ȥ9elh58~l2na6b9*DE+i S1FC$4pI Gz _h,PG3I  ڨ<2lTĜ7~p7X i2#?i,>0T|*C=!򗃭雷 )v@%W6$ڶ8,#[5Nhu^x-_!B[:jЊ7=%z9s&:>ޑ#C2:r8rނXSl=VkWT^̟mO]ZRD4ba~N+D b={i؛Bb߃ReJ]4|Z ³Y;~7?? G_q pk79]hU_D>l1|#PkFKE~U_)4&J*XLLJ2D$4^&҇H>Z ^( ,oh%cLP.! HexKJX9S : -KI 1PR<"QIVgs.{g3OiT[3V>}np:z;o3°_y0_~{=l+ɼ6׺~ '|LMUx:LOTW=uWI{{ m2 tխ+KC.6M"(]At5+z.YW%+(Y5Һ>Yޣ敒a<o~w{q ,JGЯ Ϧ'*nB~f/!Esyݕ?:f]y /Wcs՝PfNsJN٭-q[7#(z|[ `7kz~?l[!LNj͘ ;T9n #x -^gz<>̕Wg6%'g̡qzq81^P:YYE+a$s>֝Ce}P}R@sLJOrkiI"wŔ T6jҠp3-*lf`^Oݱ-Á3-5*0J$0"g1"t!#F]ɻvhˈU*ن]ko7+ {Cq8k& Nv 4ь2;J~=/i^,9d5ux,Vd8~4B$a%X\rcU^F51탞m[Q|TCZ  wύܑdp%F x3?8,xBBiIF@V-e~nPB6vڤ4Q 2@Q4;|$'Ek[g7"~|qq<;sLkCqpa6%!RH&]>k( K k#9hЌp>pXq(a-*mNp#y?Rj]MA~Q)نc҉`UɌpėy $ zֽ] ꇊ:>~*$@nDc)P8H%%eZb"#m Ӯ2Υ :$Is\jU)20ZcLH{VjiK݄& =&i냗cW{D[a}Y{~$QG>fkFF0@Hhi!*uRQ4~+cG$9$GRFq^znyШ{FDYXI*|̉ 90d@Jrl=BDi HcCXj֜IS+KSddR@ !.MZg7 BLէgiCrQY98߈gT~C1|%oQ+,hO[絶))cV U>!uۊ6$jͶsȎx~؂=ÔW o=mp)D"evg<:i;O.8F}M}оi@T&d^PI08Z+,AJKʼnG@jR:&3vZm wWgݜ5ײ{Hׇ x U:׆He)3&zHfA@+*GK2tp/r3;`!Q6阳e!kRF,HL(H@HVn!Cy9A~3繺U؁%v ,Q40K_-DF s`jD)uQqONEn:ixEg#tԐG gLd4d\K \ B:G?TLB䋃4{ZkqfOXLs~e/nD=WQoQqћ)hFS HQaI?n<y =z=7CћTK ~ꮌ_֯{|bك P)*K)W)c,׿261Pݶ;}cske~(6%.j2c|e:洔A*B:G>)R捽lj$'}p&<١œdzFYϋ_~yG0 [7#ݷKw$r917ptcoC-]BW|^j2>]? Lj*Q\Mm&BF*`r=QM/2`-j,wܲcؽ1^o\uD#ST-gZpىvGһkb]YWߓ֌Y=AMRoG3 1Ϗz:[6~zKC3ciN !Kxnmе-y'}2v\h)DX .%sv12X&ZshD%̶ΰ@ H `F}FR|ziicw*4"8!vAPsĠ^a!Q+F\"]"08HIv<C:/0^Kơ#d#pȹE^xXǫa(Cͷ϶vg,0Fȇ:ܬ3W}7R l}2-oA$q^ .l3(r &7{n}t6GZʙz\ e::e0)d4F :mQGrq=˝yVCgjjS_b~PXiԎgϗ8gu~+o4aWp9]I`Խ 8HM\|ݤ<f7&ZHVE%"%˭TǖPUs8sQ=;yݳ UPg뷀UihStm.[;y-ggޜcl6I͟T}eQԕ6w~%ݢΦߙLGq @wE4U筊<=X̗?/N p:ZU ?!.\ϳ?rk[+3yo<]qpFJ[0V&(KJ*}X@ڙd#7n/Yl)&QXM3NF{8&N )Ȩ& E\&%Hvv'lha錧@}2*Qϸ(bNYHޚJmgnK}U'W3[uyV]җooЎ9 hq(zh:GβnòޱJ`Ft#+m-+%*emqݕȩɊP$lid$e`7((9GM^)'Ce};;зs9stJy8/AL \DX"|4R9%h)ORȉK"cIsM&76 =%LZ-eٚ8{hWyYZɒBSJ?ذ'3=;󾝅o"ϷvW=;ᵬ&(3rZRRi M0Q`'P&yW^xNDܡs8cKc/nL۳p+ ÊV< :iCјJ  YYvop_ՀLPJccPֽ+SOWF" ԂWR8М`B.U*xoq n?_B7jjGӄ@bMÄjt ׎D|8+&g-S鐣.VzAyeryk@C~m.6 |i)qRG>#a9K+ T("%E~zia)P9L*.\1#@$)+5k?ޅPV ߧya.wkavZsIVZ<:pP>0WFĢ+ 5h E*3vrENεpcAl~E Y)XO] g9 m-6^\iCzZWAa0Ra; ]|V^|F^ž}9BŪps'ńdXIث6YGS+.p эhL35x^^?e;oerFr:~[n}@Cm4;M-)%!׷tm76t$7,q}8 BvbX/1qRc◺ޠ]ǿ7}&|~~{3o^ gaL pIi,Ow;p}M[Vެinqvר^qe|n^ ⷽqggQH]ٲ5 n8J̹鍋Tq+v%PJͥ 0"Rכ6ViM|DyIIMDJ.JJ)1)1B0JCb*;Z **52x>=d睝 3++m#GE]tBd1|ٙr7_mo9c9qrWԛeGe[vH"[fçzy3/1I+M6H!fkH>J}IߑSi qiӫ2gtb2gw;/sޓ'}vwzP6DܻV ;0;,o7|(Džj1ΐ b]ln~`NP⳱G8l*O؜H@:27w)B4x#\,ܞmĶ!Q5JpT#2;@x_K Ѿ;O 0O+g#[sD!l BY0sFy&(TƮl l`~`k8eoſ5lk ~1hX|@ /.%1?8Y4oc Vk#vW/JEC dP9,Śec2/n]liT5Q&Eiށ/V $*rBFC"u(PV*)8!NJ42D[5=5G}x~9/ycꕅ=ȎyNIIk  L.D4!4SZe~9$Ŧmхmf^/ς;{Qu2&$e) Q1E D@&e36PBgx:-Aڙ+ 9KBX[21Bф \kw5gK;Y^f~@Qp_i5X] -(YxB1Sʧ ڇRt0md" &>8>iu?aAcؔǒAy Ff*vZͨVOq# hSڊ(A5P`a UaP$rZfj p*Nt`o3M<{y2E}B`d##a"&1{yFnEe1/LEv5>y[XYb6|HB4h}6:Yd%5{pHIa"GvflḄy,X/?~u4Ӑnֱ+[4koxnzJ7?%ODDY$ &lYw]FE^Nk3>f:DԮ;Mi]2KW+vA"BޗQ7-CP/sQm=&V%LmNNHsdUICPA5ix=<'^1r1֗ٿojC bݝ} +?Ր-f=ճAzhzWq]_Z/O}4.,#LȳoXAf?|9; Cēn̦߾/^ {96I%$k 0? v3ߐ[ =u~ ZWDTO/)kd0ٍrM=`PTi;TI̞Ӥ=&*4)g" TXTxסd#*=M{FVp Pzua5[}W|p*IM:6d%e󸻽4 G080+yUEsW]|_[x?wʲoV{]if?S^UџWEʔWV+h]qDtluQB=I}F@NJY9 i$#\~K)eC$YJ-eқr|hl11ZkC> `7~;e{a+4l׵lNfISq~;j4+?9W{)/RF &yVE?~{٩MmϿbLrwN<gwX{{YS Z谔ܡp|"irF YЇ.vԣԨ~mP.Du6;d+,Ex "UF m@̽JƠ`QEI ){)=,) }q flBю/_{TRK\_L~:<۞lǧmWgsԪG'u 3IiE@Oɻ "P^Z QDѺW3q<uTjyh]2%6R C텇j/'^xgiևO|ΧC/8>ߘtp(kȳ<$Q:$Lс&vp(_C o%`ʝdiP%CR.Xd]F$^рuB2@56gI @z CkYB)9[*I\QyWk0h3qt(qӹa>?nW+$fIJdcR < \9jFag4lŬ )~e}H%FbAiEbcCm&՟OkKwŴ4_kj@7ӓk@f5w^__}6$PmN^O86Ve惋r_.]Nbk`Q51Y "0:@!aNtIilI jaCГcb8MJAos2(/F_Z#c3qGv\6ӌ]c! XxP,\JTSeXN?:]vB/..o.g˥/Kba)(eF뎥 PtbV3&O-#h؊TA~ X8Mα`']¾]&t)1͖ވL;\ӊJ;vEm݀}H TL쒢GeG.F "AڔeN1K"6â]E!33^tT2ְE(&Q5Fv3[kc$Um#)c?Q+ULH!f*SIZ-\1"6qD9<ԁqq͍OkʹdW\qQ8CɑpBhԠ`PpKEQ.X Ƙݨ}fڱ+Bc<]u0#] Czc.ן`MhL*̹֘.mȖ+G(C9u[1K!(4^ڒ}ɱCRq@G )RN`yb%e4))Hrs :| ֞CvpaɅ!cW[e5ra rNShuH^g9O]`P#lYlek?d+i;kD&)!tLFF2euRȺ]bq0'-*u6- &gWDC"'\ˢB*$(1e6eB$VlOX=ʛ׻׿( I8-L ”?AoFqqq8hWԯ(ٸ=Lwo'1|PZSuϾ8t4YaQm,oދ9ӢIocj)U8AHIEIi[qX$.w˝gfgyDD,pg1[r ! Sn%ǔFrdy,QVcɢ`xJ K, 8gqOp9s`|W,LrɴB0bB.T;ZX"9L |HA{}:AO.yΓӓ5D1Xʜ$XM7BKӑu6"+rEr{"]==λVg^ni/k{8Yϭ,A{J"(,ZcUydc#h CRA(p,- $#+qglWFu@v`Jp_J-^jqyW~ժM{^Y*u7]wUߎ5̒A!ewq0ENo 'Rrp^_ 2{IݓK>,h{WĜ`t H"#4bbQYTP17j<۴CGEL˺;)R|H+"D4v:g;nj#m|𼄑9a%hE }չqS)'v~ADYA \y)/_L?E]рmP9  OAs߃~#nD8ҞIj$螒$b'ijfHB,tPriu8MpSXk9#kcbg dJxtH DSDXQ׎FiupF1c1 )!R'wƔ:b(oޞ yD4ՂPKyŞicݫbZo{W=VZRNg/wa1FPE<<U<"m ض}myl[jݻQuHpS ˰(0a(x4 0b&J"ۧ\=DU5)Twʍ~v3Gr7.޿O._v!ht#= !JИ({ n3Úm(OA,B {3 ѥt@i@BG3 u&j(nI/!x-HЧ&e$ eȆ*c'nFFw% y~PKLy,6dŷZ2RWR>EıJ~ZpdpS/_XR>Q|.-ZSPW OI-6KBYCL<;G$o^}^o MO_80=j.~X4x6vܼ|C983?YMD~ f`<=y~XŐDz9P,]zNh]dVVz;}[+Jd&7*Vl˘e-Qu䖳6_ldhwc6+@v?Wඳng莳.C1FV!i#_tQ:4}vCԅsc̱|6bHsz5"S~KjiDx>Y'SN鋫_Dk.Q(v.94.G*QO EdKhܬbQvXf _&: ]d7c_)`e?T6x6vP5,~ٿ-ї,i)xoVDN^Pt2,K7 }@8<^S9D/һ.tVNJؓf߆ϙ%4.ta42\NP5`/0YQ}qܮTS9BQ9di!Ȏχ>쳲dr^32+fRJ\W"68"x wkgYevjA7`x4E O cXǰv@SL y)29"" t"uct$\~j;Wh@{Z&YC!.{[Lyɶгk<F=ֱX7`"YGH(򹖜N)bxЄ޳n܇E v[S.R(|u gy(=#:x&MlJ%BDfH |rVk/nṚk8bh)hWש\ "Ӑ-zj%'i(KaJkFYvngW?7l% \gK@4fխo{ÔV5 |2a1t[ŀZbleBSyTODsH勲J{;[Gi~tEꗜ\s&D"2WDtA=`Fiє9)NtOV)f @9@`&08ƴJE c9BW,3M Yeaygq|;sy| Tt1+ dXΦsLT snCQ".Ib{@8S9ܻc3^L1W3OVf `H1! 4:!XE&ZŨ S!dS(䩂[B7zf!% 3Bsҹñcٙ8P/Uv$& [yG7gM UhZLzTrnUD UIԝխ:Sly,2+Qt* twA!E$K;SwS-̝PrGi hFRQ""`Y)c\Y1-S WKMi:0jvpRk R5cٙ8;R/ހͅ]S(sekXdrRVZBm r6MQ&*Pti{Mrwkz49L:*w"Օ3@D2ፍ oLBi By2-3k,ŊZn9? ƵiيtHFrNF08SQi~KЈ`@~?Դ"Q#.V$Fa%# ^䉱h2*t^e"uHwbb^ʺ턵GÌZ&0`^%FA*JJ .-&7lȤll?+bDW+cp!" EIeJ6 1LXbƽO2R\Z8S%s$ă(,X; AJ tp"S̎.0y} ȊN[lB0גI֋˕ q%tWA⤢[ EE8ZNI|wHu(pMp%ou}% Q'`'҉>nR*Gӄ2"6 \q SH^#LeW.͋|TͯX]xy>;8\gk- aaGÓyRە bN\SbѹP|jkI֖_Y[3yom3mHŸQ9drl8pJVlkYA2EC#aKU'qR$>xaǠRBцJ '0ׇ?:?ǘWo| ?f6KE!uMkU^4Ul{0|wh.5#rhn1V Di? K_6MUز5?!-g*nPVR1!B*w+C tĊu#m'F|~SH"pIp ]$"%Z mC VhJWOzCcJ{>cai- B%A0e]"re)-V)Q(:uNu΢1q':Xww]󬝕Niڹ)kg Sw&uUJ&ugf _Q2V'Ҝ\(js]}O#G*Kt=%(Or']B8cM~1aVZ_WU DWT%!d}v<w.>v(>` Uҽ~<,1k,ǫkw{3mw3z{A&& & P+ծA*Awi0 BY,\ZU!WUV]B^!\iKV#"{W\ WN9 WȘUr g|BU>UĮUR^!\Y&'l-yl_P+v,TJՋ+qW/>KˡR+iU¨{Po+\v0Q+g~h6i64稊#Z˞¡UJ\sϿŰ=3LoF ӛQTii3*-`P-m0gkM0[nӼ?!M|[9pψpUt]N>ۀy~ Ҟ#ܥa`NJ=JݮW1_>5vmvVxT/`>r+!RRz[Xd `Ԩ!܆Ot'4˙H++<[Z_yBO0bB0ؽB}B]BQ\FJFusQg59 L [>u*4_J] g9u:dvzqz,/]bIɒ4NR mx88mR4͵M-G2I࠷)@Hш A9ED2*Kr\-4Zȸ7sjIP3*RM@7gDo'nzh4 ܧўF9+kf,7_D/.ek+d_7̛9y++>؆ZvJ`<+Y4,p6T'-?F+(+pWyQL?̃4]h[?òo+`#LDld*r)ɧJq69ff<8L]%.R\H\*&k3p!k @Đ Y\5r.q9 PŶ7N[H>x5%nO+eV_4'9X49ͣi No !/iM-7T4jȾ; `6zRoKnhd?soA JYLWt%.JVN F OV)A:l,]W^Z "sAFp-pX%ZV n K"Q2< VZ= m4F eJtVN<'LAA Ǹ m+AsD8)&wIQ5X~' ׿9S+2SqjKH_Ljk8J0 616gm4: P;BLV^Q4u 5eQejiQV鷝}Lv)-;{1J>y†Jn\&V{F;?훶9HnL+! Y=+$Vp2rh>gಡ$k»e${fHY)1,&j@$stږȹ[;h]S(풪*/lo@^Ÿ OG-B,+~/?B_^)lJd#̵N' 2-{bd6@VO$,!£ BkJJi&2RY,ْ/@GJ$٪9Zq }F:g7-Ŕ4prirk_ǭV 5I`IDQE]D-XǤq%.)K;$.aL"J쀧Mf'!KC 2h=@ڿ4 R* VW;%0"}ߨ쐧) g?|eqȱRO 0rO2a5{#SRD9O[,} uRߚ/%tBZXkdGCQ"j#mn&߀Ȗ6@p3hu( m ;#t1ؽ{[*aw韇Blءc]F+z#1 S]Hu|]¬ ~݈Up4 Upi4oU/fY^GkH*YߗO 5*S*M{p {/mT1{ĝF-DU/ ջyϚYjpaVI~/~?vo~(xzܙ?qmǃ8_oܸGGasHWO5izeIX\>pv>-=B$֕MHx`x<%9I5d^xxď&NhayTuYz"h6 rYYSPH:΃t60١%d_Oi08ߜk~Ja88[z{&oW/ԗb͛ޟQU#BTQqNXYwz;X5~/w~~V:FpJ-VbOW2`]gvkB@SF斜G<^'fLu>̦L$6Vr-5V軯NÆ;y}=ƓEMf#D^zړj3eQwe!;.Fcbѵs6i}buOha77/1Y"pztN:u3j4G/qӥ$Es~tzЦ8V{ƭMP*7I]3%QQ=UuVgn*Zj8Bb5h'X *!&dnat'V jav-A3H Dcn{BL"y+Lt- IӤ ʀA(ZAVi;d pq̿^1|T>\. 7\F-InPbR,͜{k-2vu̓$F{vNdv}&ՂR8q`:|8_&-b,VʗuCP!xg>R/㵿vD}Ќ=V .&wqH em/Eh89 w4kiFHWyHM4Vڀ,M7f*VcH 29&)6LI&[$"0-+QWtKP>KĄ&`$c:<Qiؒ6JT ^zFKy+_= sMa*|-|ꩁl99 @ ;M3/_b|dV)<Îevyn޺ecky%+U;UsFCkYQ!6MDօ,f£}QK4̖wb|ms^f;^VsFt2B'c@V,wbqU.Mh~>fݩEeb)ʪjgXd66M1zĠ/D,zڟJ5H:8t*IxYIs%*y5HР-ZpFꙑh!{r'c6%@hXLUkɄL~áʘC q~g5O+@6Q之A9rR}ipDL KѿfFfg!ѿhYdsՌzE^<&X2W I>YhTόU$1P2BI!PgYNa6xru0`!xz@'Y[F+(|! Zg[%6kЙ'd['谮"6fVԠ< myt5: YE$z+Ը80D b:H;!oz44(zsu1:?R3-Zͫ{?nw޷78^\x6\.WĜI9V]sRkcLjrOhɪΑlk=ì/̈&VF Y|:9ZNzb!tQ5j׳2QgՓOK^>WUOky Gӹ 9Ɲ*-tjyMmx2=Jc^>~?|s~wwoXg+.0 ?=T _wCkS/&|q;>Z8tk@Z*@odzT#Ү9WGl֚ aGy&NGZtRTvEU:ͭA=޹mclne'ɽ=YEI 6JuR"%e*@ #Z6<KmTZvS;"it ZIK0&rL2d!\hdF}^u*5qO^Cvw?L#  >Z.}v|t`h)Nr;(RbYozm˫78W^bPU{b))mEӛ8m|WOt눑c+ތ~_jv^xŝO.Ɠ;:kow..'}(- ,ϭbVlmF>j]5"|۸Q ^_T D^{ ;{cMcQ ?RCF?G|#Ğޥ{ؕM-2 $GU# E:6EHKoKR$S&LY3 kTFjy6jom.Agӓq"ӇMx$R*l X+%/ l25Vb+yUNErb%'N?8x Ji-àNdS)/X0_S l0B&yM%<ۿף`$T]QET—D,P+kN9Z$Dozp?%!p?y,&8F.lc,䄎&c$atؠ Pl(#;;Ul@[MΠav/ؿ;ާZ)Ʀ(R h 4T03T;ֺܼ˾3ѧr*rx3[\tU/;Z:w'gZAv+|ӇddQ$ f“AɁR>[YeR!bx7Im X9io͟`<$=ΏFi2vU?~b񟩍8n~'sz&'OJȻC;p`ü˖ LO'=_ՂZg6y͖ @*ݹ!UsR~N~I)4`miP( e1G*;hٻ}bGv/ek5`4I"-QZfڔT݂l!vYB2 a[LBicڄ},1}i VQ&)dSQ1:Q:KE{#ngK憄ާH(>-GΚ5s<-=?j;ie}](& 3ALPV!&3{{ \lFPbL Ѹ&j M1/Cir`#e2e$E1F2ZQ$#Ġ*xMf|HJ{=t%HJن("iM>FRNx)]4!X"R;/rFU[HDz~Ăk{؆jJ c*;NJwg hx/!5F+$C[&e')[@ͬsz<쵎VlmLYR48fQ1+%o] ޲:- Y%h^H^$\T U@yx6%U D i$-=Yog{V)lN޲PDL-=eabr:dA&RL\ʗ 4'5@,KY A~v@erDXg( Rmx,yfRj`' v>@O&.gaHKB9+ e s,JrAo F`K!Ap"8@Ookx5"XK !deAd`Bd0ޮajiz1پ0_oK{fUP̯YY~p@!Y"P2*Yؘqp`cƚ4=m܋;n@y8w9A/gUp:';&MۤY4y EZ QN C36RXͦ'LkO+_ju4$09Q) ! g4DeTL#B_FXzz}!o1 8A9g- uls3db.E3=:]72Ȫ uҍ<(x+XBd͍4 Y?[Ry}Bc0gstG!^UM 54l&Z`@k }5@@{I?y׃-d#c+qO2YS{ML)PPp7uѲ󟙜x ܺvIuoUGK[ 7SlX>uvyt%zrr3#3Jb!qDUiY!E4Zתy,܀^BZ}\C c ՔnWVMi]H뙐_;:<ۯ翚&8e7Y2NwRl7dl@zI8Oz{<~܋IdKm3R뒄V[W V *kezgr*(M>8G) F(G&-b +R,Ck WTg\oX|.~\_E6;&gW\O:[;t:w#WjvAM%@ϊjZxgmap|s+[IMGj#PkqvdO*u{]uí篭ytNfwtjsw^l; lunM=_7ʼn;rz8N}wYI-XW8:NO[\WKs˦nvoq)w>QPnپR_lspӐ/duȭ _q EgTP&MFLhУCvl}Ud\.E-<@%J`N :&J(>&ou4&PL*ZbX8_vU$1 PvmRsv>|wF_[c{w|n0>&\^۬Dgf}!X% L%iQԔ@:'҅ 5Cg${.^4%b ޒz$ɛ0IƂIa S$3UI<ȹ7|/]{_- 6qM17%|&{#<7hIz'̚\L0(kZ^eQ@ LѨs=ç!jP*o1(EKuI*E)k&'`` +zN :N*UT!X@ }owg\{Ʊ4Wc#`w'bE|IoC.K{|)]XV>-K/yȇPd|M@m~ Ƿ

ƸFM:Zh8# #϶`Heęf6M|ӤٕShˈU)Ur K0 E D3.1TfĸWN5%GłpKzFC'6k &pml$4Q?uϨ,eZK #He08 xxs$h xil58Q Nm4,GΤY+7A$̈́my_?{0u \\ltiV/i\={.{RIf ":$D dkx7f:Jō'TkJ3aTcp+G|e/豲wMnӌ4fXʼnIm`ʈ๲O{aZ a(2l+V1 ?qy(vnf _ك%/%b 6;H͠&J0Wb`v7+|P}Fd!|w[\zow|wf0TaCrw돊_,Fh}曟~7|Xw^{ڮ ({\{=φl5I<6# ad@.W(WPpjUq*uSĕJIPQ Z @B-z\ 45+,u0BJ+Tuq*qu2Zj 4\pATmq*qurP `ATkIq*U= xE!!KǺ:ȥ%^|ӔiUMht]bp"G''|\_AmN8ȘodDБ4C5qkm똮#Xcς?zrjӵRBt-na41酑%LIeR=Z$,Kqj Y,(_zA)mfr>R %p"^'yrQ.e/ ~m(BYOoSkGMl6v /$qe:DlBxW =DE \1)B#BpCkTr K6_&~ 9?ڜ ;!m)M}T0qUz (OOo"=}OxzycልS8ga +ʇYҡV9; G_Dbu.5[M_'|vMQ4K2Ql:ç8C$J +S@Bnל peqQъQZj*h~U~Z fدc%E !E pjE)΋^~ؐ +e\FeC"φћ5:Jm/?o0(Z_kmz~GkϪYϢlG%%<$_ 4ܶSke}5PII﫝$3&$\`d0BmS+Yq*quRX0 Pn8Bm-^R:E\ 9 p++E(B =NWRBhE 6 Z @-Qe)òJ)mHsW Xr W Z6pGV `prE0sWVӮ US'+`, \`N0ru0mtZ*%a=NWV89\YT< Bz2M/qi)1vni(U#F F1CKCk[%}:UhB9JV̱_KSs:UZ!NUKJ'ɕ-Զ5XOk\Y+qk f+Tju\J#z\ VZ]`˂5+TK;+TXVP$ * P]4 JJ;!y='W+TTq*quK@"*\\jBQ=NW Āpm8oQPpj9:Peĕ6h@0xW(Wfju\Jqe,vy3r5 Wt~ U~qe-54+JSB2+TIqU= TEӫ!!ҖF.*L7lDH!Q#݇(1;s未(7_tj6pNwN5Gy6bkoPOnH==ym뀖Z*YhQ5^ջZE;p}-`pr9WVtWRW'+fd" \`Pf ՚ UvmG ːAbZr W U2q% 54`:\ P]Ԭ ,eHB+kT(BVwW@JIW(X`prWҮPi{\" YC:5f"iEZh3I!ޭldS4׋G.VR+4V-m}#tfW?GO6ꭿg+\K\ܣb# ~ym_Zj,k>`zts.~` )KBD27jd<ŹIc8--lv8N:,:yhJ0ً*ow<ڌ}&S<ɔOeMd^We2W];qʧ VA{쾳'{`%^jpWEunxU7͖K~;1W Zk@[hK^)80"y2=Kzmf fq r &]YMXrz3ĤZ BpB[ |dD^):sNV:'Es7!iCB료Cg8TvnϒP0M*hh*mn|$3Vc >E>WC7#N3ŸΈs->2Қ=li0S\T(SJO>P]zʘwGRm޵oarSJ:ƪ`ZzrmlRKIKK멤gK1v@”vh0_M!qQ.]lwΊmp+g Z[s*{kJs{cR&;a-`\BvS=m%1SIKȨ`F+i8,e<\ZU qũ+, 6\ZyPiz\ R0W W P-]U:A\IjգfE+9W(WPpjQB:A\)FW pAkjt޻*T)J3NB]\ P< =NWI"BmE*+BoQeN9q(LI ,\\UBϻz:2MohUzֱS`dQ3bjŐJ޹bFZ5m۫'US[:;ʮ囙6=6縋?LT\RR(En(=rc|V nxq;lvZ o{*5H4>gqbA:JHF};AJ-7 _V: F )ɅO7ƿ<K=$d|Ue U_Yᆱ]M{\muƸkhoˎ;dzы0${=*1}C+x8)Hs$ZKK9>Fne|pAs𪓅3b 1$1@?&!2eM|Fr)Sg{㸑W, 0^'{6n3,؋!iyԣòn~~CF3-kQ?"*$AUI !=#:j'ƿ4k['feOLN]w 7 QCkzg3dҴ]rQ6pNVBҋٽr{Mv]"c?obNe>Xֆ\*M_ |S$97bzz4˜{]k-4Z.GͱOÏéu -߈_7ga'T ZJ]ONw.QhpCbTl]s|>|bTiG5=$$Q]@NZV5d)rEn{jr+mg7 j2}B?zf_2g)fjk8aYH >IbV/ F| G8|D#FZ؄S LBû_@LbVkMnjm%LxEB!] T^\j+hY(Q-ECA ZhH ):D>#g!ݞ=*x ^l/fOmqgBV)'=9'HM\@L)J .IJ"'NQ&Q;&6F \h8Fǹ*D&F64CevF(s] \蕽B+ /ФTA\PqK]~ nn>ߺ KG[>ZvFg仦ѻd."L K蚐`qBPJeq CS:^(heŋHdXq<*2%OD"N(:S @? 4%"Zc=({l2za]]_QGS TFy!EӂW"jаdׄKE;g,C g'a s2M0SY^|h$($bЖf]p !WNԱ UX=Uږ߉Rtr.du8'QFh!|pUR-%/Ԥ0)iX iR`+*NM=Cn>:4{79#puSD$HMY!?aze}Y=L}jhT1*"gI6ݠp5qiN'AªtvrENk9NҍnN3T'qd^/(@J8j=!h"1kiO4JoV?oeg.Rbv>FM:e!V%oM{&:;[FKg d'NgiuOamfL/6V\ nT9ByzP?kkb>h.\g AE̥4,h[ns AV+gE^yab\'Ŷ$.uöwnV- bQl覝pv@l핑{mmcE^9[Kd'U|+ r$ǰQ*ԍ .ߞW?o_/_28חo_@'Y}69"|qk횃wWn&|~9r[ >z+v ~zQS?7/ǗWجzŖ\N9e$rw7MTT6UBՇSVT\@ XRrg՘|;}u4 ?m67' lD5H('r1QA)t H9罟+*c :|tޖk]סQh**G6R N`v+ ~]uiͯdj&}p+X5# 띷 s 0ƒPyʤlڹ5蜵V;Oi2R4 nQ2qOl }(-XD DP_:rPBDR%Yic`c9ֽO8Πy 3Ș#A7+Rdym9U3A<5> $)-pr}T[JI QVAK=8|U$GbFB>Ǥs1 %": t6M2PmѾ${Us2BZ#PFKQ%4D#aӅ(LqE$x/cG$cX/ deon X9p`Qо/q~Y˭^nVJ%‹b6.ywH,Bp (1+aʾ,ņgѱ ɡ <3\gyYs+%ş})دDQHvz"\CL$]ԂygLP^*g.S"VЄEE"(T h~\_r0}Fbx2\9|oC9.ff7ݟC\ 5p %hkS;.Ck Zg6Ѻ.TO(_aE%d%tvJT**WNq*$ x&6ң[;Gv*-F$j#q:9O$'5Ҁ5&͉ED8<2nt\A;S"hAl=S9EpØfQJ 4mA]sL-#hCeh_)n-: FVTnn(ν%q!@9i|2xe1!E3.9Ȝi JG8zLrۣ= ;ţA))q!IU4J;(4 Gb4ʙ%r8SPdHAhǔ&, fP6+Y $1\qkxr9;ّ_\;o2dZB &4(JĐ!?F]TҺSkOIV& NEAO?13=#Q-ѨHx/Xq|ΠP%DP3gO/pvDSuDj⩡$')Z@{i."Xu" Xw@\TD@Jh Z 0%zpSfV3^<+X3K:7.`A4VoGnXnn)Goa$r&J'#XBX3A2wP%1y }RI[9Hfx3n9(NLYb_'HABf_?fVSw={"G!sw9&6^L.0$e`,,b(,"xR?+jQ1x,>QB*E[Tx1qWEӡd+J j!Sp!ZpT(fxM<%o+'Crl0c٠kmm=8h 3~q,)H)C21)#7:&ܫ͚+g#hvO\k{$7ޅy"qȣEB=/\UTB.LpmeDǸKYpJH2jNYV:3٧vO{;͗ӓrf<,H? 0˳R0OP:Igŧ`Xp8zl$RpFKRM3!t}SF댴x9p\}5>hģ-lFWV\p9hh??'#JRH@ Z3_@J>SgP8ʉ{|ʉpDʉ-h$cLP. !˴a8CHJ8S|ܫQ0.mJ&PcpH ޺}ND Dy &g &osLud9;4GZ\-(gv0Ց&W7wBOۯ;4lr3"{͌3{OӍ޽լjsC8ܕ;IX,lBw7[$gZ{_ӽL FR$ٓ }-[L۲2Ȯ:.Vm~I+l]?ۻvtzfw&Yw-+lYںۦFWwwn R0h՘(wjoꎎېf,zK\ t]fj/4?ӧjvGq[fn7$n.TŹ3[캊lypԖxJ+A ƀH;q#(癗N%:Z!MyDdhCTTz*/,Ns6O+tp qs`w-u3Ci?gaq6/{ g{0 8I&-A\RiM)T;G 9'vZ(ʲ`LxgU*kE3Jp#3):5cwhT Hamkg`ǙM.=Y6渞S4Tr3#1K)fׅиaΪ_磳2\Y Ɗ LR1]l_P2~tYʃd)O(ʘcPSb[IR wbJSPE 4T2 }Dikp^oyqKJhB C!qHLj$C*䙌] ѻ=іQ.. ӎCCv<8$^?tۂG~T2yynT!KVb8%BNU3W_WHIGBVP?T~xWMH&B%T&i)R hnvX0eÆq.pb IJR$J9匐'c 15ϴ҄'VUK)qv;Q48f=Tlɭ0>{rLq y$GSխ?<#3c?ˇ9Kpeq L63ymJJ+X^ f9:-bL [sm\n#EjVؕlkێ<1\cPۗ ^ZJ/3LyJq6n+HI Rp`:>sOՑ'HSG Mޡ~rx822)AyjS=%O5d S#eI,%D[.!}$cra>mvԩY7p+~mT6?C2=_-92LT1s\"̘"!p͂Z,G4L:DDȺܪq 0<0$*HM:y\ F+e$Qɢ( $D*mH;XWGz1CO(1>ݭlfr5d'Ath 9CW454Qo:'wU笸''E" tSxYg#tܐG gLdz\K \ zEQɩ޹(4wSy`ZnǫGF'oXgrd~ = M51p48+t4:j Y}1X~lv_㐝]B)*~ Y_*ـF7 mg 91q )Y2G8CUDMPf,cb1h㷇CN{N+ާx]NNJ!;pVBlKe5% 'hO9+!䰦Q/F{sA_7Ƴf,{)JJ%D5NiHnwZiz8?i:Laս/FwoFiQ uQ FP-y)ׯ{oX pWSd"h>{_OkXrچߡ?uM~n ?kL}S79z6#J^Vnr9a'vr}}e")O":u쾾)ty3dSßnvx<%!#5٫JW$Iڟ҅|4 #mYr髭wFʱ.`k),PEj-lnޑ3l3޶k51M̺Ӗ-X~{ú[N=,DMnSM5AW}Jvo֐S[a"VGӹ 69'Lu%Qj 42͙"{# &2/Q9l脠ځsĠ] z?.hňK4*HI ʀ^@HUx,7ơU+dw8]]/`\I|>T  Քn Tk 3%k$D']"bITW /^tyꥲݞn-w}%*q{q'qa3z0Bu\X!Ek\E7S1OwRqٿX>;ƻ+r~~ݤU~]dJfqvQKrj}KBmg"%GM^J⋖un( >: mrm_f/E^t{ Тո7G_7MۺkjwѬ>ޠ՛o{Fahk&7uiei|>5gF\Zgo/s[4] *Mo:1kE.u^f6[ʦ-eٸ-׻\Iq!n+O@ !Wpqra5A+nqٖbiC=& ?W&rwg{pf1molninŤ.mf+h{#I[%pI{GƚU4w(] \I}bVWe 5o&~˹Y?<5s0|w1PQmtLdqΤI'NmҥW:|S(2ܣ'%VF*\$$Y6QJJCEJ1cL)ڌ>f, R =:A\&V;C)*~֛M6  ͭgPƱ+ 4G$q&P;szQ!}C9G܅Nk pJψ)1Zmj!;f HK8?8 !@riE5Qz'^XPOm Jqs4[PMOBŤ;$R|$qkHR A.Y#޵u$_!`|f<^40,f놑O[L"m( V+^nZ,V'3#Oț'3z?u}2:5&M;Yn]=Q͑OCZG7H}m fQU>6$ VRjH) AF5LIZuC)hK -掂|^UbT 9>tF 9yR 37g ,(R>` :`՞JD{%#vtP4BZ%{ l$S օܢ i h**u(:ݡ-!xy9xi@yۏ+H%eêaD]v7C/qWE b:cm̭l+.tlNc[pf4giF}/+g) Һ1 [H6blPѦ"jB-r  \z ,~/H1Ű&؁~ok:Ye; RcN)W̙l`_Q&t$v xiM +Q|!+-"w@i|U Tӌ$߈qvL6B@F=5(!ȮdIc@CnTzC{- ȸCLAA؎[@zMk\/b΂G ݄b jS]w \t͇*QKU r 6lC@WJokLEwR )tiC8BpSH( :_B4KTXż#՝ mG5juzui`M03P;v)Ru+`+tmkGk%o&%6h /VsP6DR& +epk ) 1 EG"<ѫf *5Q:bkVcCX DA5D/!PTC+P:m1Ѫ2aaaGyB4ΈIe#[qk6lc:LNx2$?< B#jw7dj8  F9K` Q84e_峤^y~z 1͛ȹ+LY0:yՃdllfѳeB(YBHu:5WZsLڌQg5rFCo bL P.eѠ֫ʌ4IMG^AIؽ{R˩4dsCP&fc_ >"$?Tur7p-w$UycaUèy~Ba E5FHJw:Tg=s g5(:W!_eKULmTƀ洁~heom-WjPc Rת%_fs\=L:* Tk>Q?mAYo5kPqab2µG],Z ,U!:ZSDàegJ {̀|12 =pe6\HퟟhJ7a#5>dNҫ֞k(5@1_5fs˥ZpUn\CC,Z*cQ4k&ɔZ&T$bqBeN V3f !KkBE9OXD;ע`|JُL0H5zF]맽Xg.w.K;tcb tɈ|2qv!so:eH:44Ow.~o6{A_A@l[Fy ·|* G]1\;M1hQ9U)&8? ]1\f+F6_ 2Jb*FpEthS:]EnUb/pdf+F ('_]g}zT)mސa` oElokᛛmof߯쥲_._޽7ap?C;w*^hui%$ӟntz*{~}nP|?skޣx oj w{10_\_a߼z&?^]ܾ7}x~d꫗4<-=Ucn|FݾXЮ>h=8m~?wM*x.]UN/7o ߻0owg ^߼bSHCer~yj /ِdd?Z ]K Yv)U0z#^@K|<}ptb) 9q =Q ےiN"?uu s:YWYњub*tu>te2*LDW _W ft( ]!]gODW i~rbBWgIW>DtiyYўjO2䄮ΐ(CNCW WOS 2ZC[+FiΑ|D~"bi+jbzQnr/BWQ)Rf"bMCW ,th?a*/+Fy4[|芟Idn+F7_ 2ʐ"'JZ؜,Ņpɞ$tƞ z>]ݽWӆ>MsL3BK/ɽ6oH)sL-it07.Vvs:J[ス ) hFڍ* orp_,|gߟ}~to߿78vW(Nc7܋9qҴ);cz9.GdJ5Iz@jN${ ^,ў({(d=]p2T uateTR: ]!]YG/4t^M6-DieQq"rکDjE5zt(HWL M<ZRWWR;3+3+4tp4thQ+; ]}5t9@ln%qQvkM:9M/NMm:-PnM4?y]Z$I}]}خ%m'ld>aIO$ ؟pO|qBo]2ʍ+HBc݀cj`wr2BWQ:C&}C֣+4tpC>tA)3+/qztŀcj= ]1ZNW3+.L,)3b6BW6_ 2J/3ϑ ʹ܏Q~bn JBWCW&gZj`{r 2avF{* "N3t2a"bMCW 7,thOW5ǔ1v"bi k4OAF)"2zZi62l/>/}=y**uL-hN*kv-BmhzʭfдEf.Yւ o2tpÒ:Z/n-4h8Ή.8{>iLZ& 8ijegTl"QDtŀMCW OCW6QF+tutnm<]1dΐ7U1ȀO~2^BW6e:C"JDtŀ< pҳ՛=(tut'+QWxR;4Z *x7p2OCW ,thikWR9ULtŀ'+,thiXs䃋35Ӭ]1Z+Fz>sknItӥ.6(TGQ b~,ḱ-;J6NY.hY2IG炖%}7ë^|[a^= #3STCTnF@*QsĤuozS&/͑1{b!|l_ﯮb@}G?r}c\_@jAÂzo>yxU۷)SdΕw۴˶]Wٻ6$W}s7=`'qch7i>tJRJ7WDR%$Oқ~*p%Ϯp \J VĥUC$h[+&Æ;?J2T*IšU]}pť)Ax}SpeUNnMJZJ(Sr'~2pO]C$%-\}p%b]Ai gERH e穿a£eח.mT!; ~:{f=MW]}GȉR4gKr ,өޗgn (<](d]mU+i↯oCe}_xQ@f*U /ﰵz6'WOâk(o:b-Sx3B\Fl]/z.y] ;u)Nnђa^IFTQ9h0I9DF&}DbQh2 9g}xc!S܇ЍsA9N.zsL(QmGs+:2,"1rPJH?@D4UaIcƥϵfi \EWqM`.*A֎պ]Nٗ8RӳS;K^额G|ڒitXN7Jt.:Me<\>&v͸ -ZPvWssA.jXQ?Zc)f>_fQ)~ݖe(T 4( jғ 0"5Ab=?M8N`6&E52oAN'Q  {Z2څp:yQ ø/SSzٟXXv_'E_Ivk"׳7<Q@cv4J3+QH :EU K·*o#;#_N\w4y~vk [Yc[mk=IZS-l&og3ҶȂكX 5KՔDR6%}ZD<-{ De/Ś93.mo7mX[B90`ZE Ä` F.$(8ˆ([F1aR/13)⃉KM`qkA #(=\ avܘ87jq9^T!QeUǠo3z#b^o U=:'k[U^'g/s:xĂ`CR!.4&Jyk̰f`EJ`t-" "<-0] $t)O3QVRőǔY„.Z5w#GZam`=w6x4%0ywzSs[jlof 8?g=\]_ns6 I:SHq8 8#Xa#zlh?%_T/8W&MzΫ!y37R"/鯙4)z:Z't@qwϔxE}<;QqO"d3*¨HiBe[$S|!$E5W姤fIe/{Qo&eCOnY2;3:dFw? {]}\0ܻ|]d6VdiIrrՉL4*ZhY*yWgx/NJʳ8S;t*GƒEquZzTأp2tA0齝\.\]$àA1,;CU1;r0pןv?Ը!`{sơcd:쳲Ą\K3mn-V{1`~-&A/q:;0F ֛Ⳋ,{nPպyz/f1o*iʪduŖͭ &v;z`|I&Sg_rFMHݾ\f۴;k[~T@L`b4"eHqW^+l Qak9ӈ<`"5@T7ư4a1H^'gRd"yo 4]|W>;F}W/xcsc n_΃t67](wɮn 'r¸[bS v#,A{”J"(,ZcڃYW$XWߐýQս۩ɶnimDz]d89H&<@~E>ג)E ,گw9ǁJ饝OR(h؇o#;%x϶jHlJ3t@HmojE:;^sp`F"kj; |s.WneQ4#_e߆Np{u4Z|( IuWnᔛ(ܙ!ma[L%zV g.-+'no)P*)k.T(|r8Hs6. L,ɩε1gB\!"s|HgcmPKѓT*ῗ,;C*Ō4(VX:G4F"]<0YTJ."o6} t_ ɋ%bAڧY>=i]䣥D1yY!. ~9_lә`e9)܆3Dn]9"p,bpyp%{}L},嶤LJq0.F-UmްJ`.YĴ|*C_}I̓z=KUW3XBp\IQ>6_N_8>U+4iåp|q75w_A?.p'q1Fؿ^x7i=w|wnRҥW? E%HG$Uw,,LKIϪUxEbfG=8P…3`*u:J7׭wݠ_u'/|S.7#aݫav^֗o?OJ^ 󁚗f3 Vɒ;)T-K fcv1.]ї^7o/N/9w_|6<0[aW}{mh{t-;,^KդamMR]j`}5Ʈ`yGhb`2GgW/65Zmuuk+Z:-i#bEI(E_N l\/`<ۙ ǨPlCo†A~@OӯoziE :[%|UU[[joQ@y4jz[+x&1+ n aXέ QR |~NC9~nҖլ9ۮNe֍ .BbDAJRIRp[AJom{xg$F&Rv%g3@8t c֣d:jME|B{ayM=C#:NC2:+ ">G楋Q'A;􀌮}*:[mL9~ctzݝ;^5N] Ux"㊫øYa^^əzoK|F515 #Lhc;:̃;N";tD<.&FSz*;&Ed -8 Ў;u$xYK`F'9Y^FXaY)!2ug(ʕI_5q'} lzWn. UN,Pیk v-r1`P1sʐY1P֤YIZ"c^I/6?ڋĹ P9h~6h5V;SCNk˔IA^pLW6dkY6Z8fB ޵!(H>^si"^^ E*jC2]C!F6 [ouXԝE>p\l l.bh]cvߛeїL(K3t{iGe4VLu;P)HM1{ʊF,Í'lIR5Y:Z8Εt@62VgxKոJ5,62vBcAp)Qieak\)g7ewWpq6]wv6|6|d%0aƹ $JrND.)gLgCF*ގ'D YR7cE`{ 6ZzM8df &iu)`2bW~QfbvڶGnz:02I!C.ghB Y0Yi\L̐KH/hHAqL%s>f@Fu"mĹ[Č c*D6>vDD[G7 3pLhO >.ٔsHِ~+6`T]k1MՑ6kE4r}@t܈!% gʈXMD.Nrui(ye\=.1.Z&N{}9cώi š;QӐ_$dž@4]4_ȐWK26F&sSfcqfыn巎>QvZAq g̨. M:TD{p}=L뉘aLbWS[9t 4HzT`.JL=6#jXCX $Ņȡ8ɊX[ndBpL}Ġ֙h;\/ж=Eg틮rhq'WKqĠ!~Օnf;G{ ]>sbHO6nޯJ^=|%ܢkOnxCy#ڼ͘* ~O=R^kDˡhm"]bbc|9n<+rm >".| ~>Ǯȭs:5BD42Hf}uML -fQMxQRqfv3OFY3fanGs먫.O,N ߉d\6~nwya5E]&J< XF,\Jù8Ǵ%ŒqFWu=*{dy 8 &CgOg 99n Z8RXPAĹ_Y@%[;Z-M6̔\oWYZҙg: VƓX@;1iA‹ke\y[7}w+t'W{ ,kle+0 >8` `oۯoLG#&4j84<'OzFշwo,ݗ epL(_Tr_Q^xjfާaR7lC۷ǯZ`E9"q85^BLz'4Zhܔ3yЩ~V{- xv,R=9}>L\OK{&I: )ݢWwW-GXWE`{\Rp-U+H``WEJzέLE0- ` ,E_q IJ3}y5⩩5l$̐귻z:\ U `\Z!+T)i ĕo2a=\`&h1B+YF:j}0xMdIoAf\\^P;UBZ@\"jW(XbprU1cWtWJqӮtI D+TxqUqJXbAn?H[u75gvWJvf*VTtWd2UwWGqjo^3Ӗܷ7Z)kHJ(!Dlr? _#_tҢŶZb4x_eX%8RTpwWg3Gi-!'b_M)1D0DcQ.XaTX׭0䤷hzDp*W Q P-|*quh +KU)BLuWrmt[mJ A0(WPLWRE.WJ3 ֤\ryjOB/W*ctA3vr,Ww]ʵU=.W)HALYUŌ]Zm+Tizwu2 <(W( m W(W3ժοD@\YiI ZP;e u\J){\}3R&+K;.d r{eBAu Bc`JϽ)[33f*UTLG=|l:xR;जԁKh7֟4/;FK&DA}fr)ZNXDǸV+U1BB+T{=y=.W" 5U)BUW+!+ W S P ju\ʵ9=.W^p9 bpj:j}0xRBQ  V+BtWk9{\ 0\p--W\[ @ B@\I.]`d1B ZK:?U~qe%J+lH9oQ.-Ww~oWׄ sPvE-n_྅1g\gt7cyE1g\4sXKIsT^:F93@m\[7S;Qn`uo-z*V \`sM.e\Zy\J!z\] q \\FB@\qmo"XƋ\Z!+T)Y ĕP*@0;i2] P-]J*#eIcW -h W(RpjOm]_+T{wuR(K >- rY PD\i͙b rAI1 =z*m ĕђwz_~ܦ¬Q+~?~A82Lye,wVz=SŬ`r?a8rk;W,vbtj!ӧ ?}{`vQ]?Ç#~ݭO܏s/Fn:Ƶ!tVtYjs}_Bq09 Oxċu_֋^:Oxj[⻃E6MҼ$g~0k QC+H]?wgs4ݓ*f8f! eg)#伖rw};[~=e;zyp+|OƋPsEa:u׭{؃<n5 Eu]8A1<xޓo\p!|JnrBf23rT*KDM$*\&ؽ59Bu?-@羹 "^v}$;nzBm|{Ӭt?Mgm9`O=AEf 8wX ĥҚRvG i TV[pTfo*YJL28ll3995c2(Q@\-59[F c bW97Pr{\Gg,ֆ i_]3Y$Ɗ |& }@y%] (N@yJe1e(8YI軔N"AwŔ PE4-7]4޴}o6?-l1nQPА|P(TF#C8F@0b8hK1yjCMhˈU*hن$wIo@@H4HPeZZnP{\QM3^>뙕(\#\eI55*u(z{}ŝ|_܄&d4r5L!j--%I4DIϽQDmJ^zY4LRQc.SI_-frNR)\HM'c{^5V3eigIYlfI/42 &DxвynPB\D)8Yiʣ,#g|'DLh4\2[.Nkb4Z-9e.Ҟ=wpa6%!RH&( KM"*Y5SJh x .vՎcyZ!;'@X x M-5 xُώ|8V+rʵ걙Zγc3wkG`Goxt[?d{z!O3Km]7O6h'ӾzWQ Estz߶):N5erN/Unu@kF >9KS2Q uiEE{SN1z,;F=xυrTa"Z2dXnk])Jgbx3I A_p 32˙F\BYma4^ gu$D3""+R4h6D4/.U.29BZ;kqQ"CT46ioJb"A A11o{V[r7, :͇<~}FP.w|c8ؤ6à &ܚ3j^Tucsy%@}2w1TA~;|Ӭ \h6s+s=10:g$N5{޳1cy,Xe{U 6mѳ?MV_/wi`!蕇fOW1]] Gt}uk oeіF>qsk6ſp8iyj^Ѫ~">3\/;iiQkmiGk? [SRb\%5HVX%3\h3Uϖ` 'ԾTv ́S} J*B[d^P 2L9ќ,$DW~S^wJ]TǍB.p,Y$B}-9àމ|}Ƥ4nQHɆ`S <d BK׆He)3&zHd XIJlQ([0@lM]տ#!ܗ0{b8yծ2jB, T#& TI(MRӪ[ڵR1|'Np~8D⼋ C{B \ldONw 0{*E8Iz|{F@ |ąsc̱|6cI7>Ԥ*:B' 0_k웞Hbt>ISmTTèH9K҂6PYj0nc0;$ӟ?{?ԇ$/{G0CS~SkFo̸EMG_ッ}vat;.cꃕÿ%,ҿwp>M+z4.ΓXf쩼%yOAcfzu`VE/ta >t<*<^w9 N[O62 d'ȍa4sc<I8jfh:rQc:L?j,?g$|Pɕ$|~#D]1./)`S]0s#|>ӧٟ*|Zlskw@b6I>t0:Nnh ?44}y6ŏw?kXE޼IfAz-Jߨl%w?Ye#HiV&n᝺v^&toru|41ӣףwj@NlC;sF4*So&y1z4 3xejj-5_ t^{7O_ki^o2IunE҆ zy_'u0_y'&5'{hN7tGxk9gHDa:539jox+~T@L`b42Uc!ŝ^ydJ/8NEFKv@3ا;L3QK'j-x|gKt+a Ku2x8/A& H?t2GD4z;wv^Z.~K⮔fhmmܶ!o״ w^V(\PR0e{RD@ADs$+6+)B.l!sy1&_U Lhel`92Nrkt|'wYPh5D1Xʜ$XM7BK"\QH#g]X6 wHVf:пj8[r`A<VrAeI*p䕎[:- uUNy3GVYwvSzf]Hw*D,147ClXU/I9Q Ul"H@ypeVcUWyye^IFRcEU+bE1$2AA,F Qi EVJqTQܦ:2," #aB)@)$m49WRhco qλm!%_ v灠ҭAjuw 3_RbrWw +%x88g c: $ ,2) ! l3ic8IPzEsNgf d7~A -)bJcR(CN@we!v 푔ͪGS5XrZ\լؕkm<~ÙҼuBlDlzoA䪙 Ji1orrDY٦B첃 ۵Zlw\vnEK2Yx@R`=".p'C(G`)AE8b IBϡ=%䐤2@HMLc1>xMrhFՀox0 ttrmTNO Kx4 /w6ǧ壯1ʉT!x "e&pfJ(v!a  m Oͅ0jv9RʃpP`zqy2F5=.C]sPbTSPB])F**aH: [paՄPg4r뻀 :q YLzp Q]Fob3V+h5_ #lGn;:@bKYWP™FjYP7U/<)>}Rȥ[>y W^y7x49v,|F.8^=oڅnTV+(̀/JD]@|V+,q?KDW'\(^p!/>L)mַ{++{a$3lylٸzRUὺpyP,:XEk6.;V·do_e.^֛_<1^`5fdIjJryuWo#}LgJ9v&m"شsP@>6%rϑ+f SWwɭfl[j XcU̓q:1KA)nLO3{am[  8` H%]^O.W/0!/~ZE%1y=u3ᤴP ؔ%٧OJIYKyxʿUk].nhPFrh#B¤ʕa2=cijyY6<ώ bcǠjn>sXjś?82rrxjIƞ_qS79X#PJ3f1bxYl*/ zm+?u,a#cĥGϧ b,xgIC E*`M|6( G/_yeً}LT`,DDw&;\k[tMkXOmkKqCnw 3A\*[K0{3YUvE8xy[v'cS/dW>ƐUhwXSPNJc],=y$Oc&y4Bi1֊:Jl78̳3k4{@r[}Ե DA 2$^gD9-uNAZl,1EJ:sO_K<ǿvtF/"jm8̥Ǻx_F¸\EZII']I u{'\v%Z,L4eTgʪWաG)c8eWl[\W)Ӊ pсbrIRfq:hs*ڒIm$RȞwDm-~:`ߞK"Ms}X5r1NFK".H,#h ):;_(K'j%YzءsɎ"A9M{fLJ{b7A| k+//^Ax-_XN/H&b'ovxIGՏW?VD_;L9$iqd/Pk_Hqzf`]9J^%"Qskә^4(t"=7ܥNw$\P9ENTzJj&};=VM տYo\>y\t: QMAR/y>;.fWRv_h eqtItIBu|EW@Ν)ҕ3Z9^ ]-BW s2+MB `;^t:Zh}8wZ(Y"]e /8ˡ+b h7G)_?@״VZVt,˗DYR+{O([ ⃿]oAsyy|RU][Ï$Dǥk?<}"QsAoa/J{o>sJSP싷d| / ڧ=>z]]ͳBLdD7T,N 3#PQG^H>}`><<!<,jjIǗ.7:+EIW y/QZY;09\H܋ 5K`U!+,x)g4 kˆ=Egh͛Yovf=\)Qv3,#["\}PBq=뚓^ےߟzRťsCJc҉yKݷ~2؝H=ڕkzEaLwcRG@PƜr2X \7JN|'8F\cNd89Ndz2k\2 T6bh^ biN֫sr%pR9b{tɌihƜBpm~^xN4)VKAcX?{;"!%QZ4%М'Z&G\{m茌(xF4 _S!d}.1(&Ykly"16{JcH\9 )QpǨh{Ϳƻ^6>dJ1|ST":WDyKZ4Z%0A)m=׎yy8HN Vƌ]Wrړ[v:1%-ob2XJY[I9Q {d>LlTfc=#`O.8޵.6nP_M5xŒ=f{n0B` :rӣxTGBЂKn3fT"eūP4AZ]rؤ)+dsc[ ď&]౭8zq0HPRPyU 550kS {Ah#&qWݛ"uW6X C+Ȇ.πQkK2/H DF@H[ơMEiߘ ZiYC*7D@K-iջѡqU\B,-\bNqO֤mjgpT'`2 I29M(p} F4.yO +z>HjO(\q+Z :wH: 8/(&`'*zC AveV܁uAmTzCߕFCd@ M&jZT^ߐeh"K^@JJ~@ 1|H sANTm~tXtuqW% 95Q= -;!>doMЗ&BtD6X!w<~w u *Sm.VU_}+tmHe z<ԥfݾb>MzC*-]%@_141R~V:@R@"hʠvoZ%BܶKAG+Danhiǎ@H;^EwHԴ^vm)뮚Q b5Dк8(z yLgB Hdݔ5 ic vٕ$ fzG֍Z. u"NtDwp{w Xt>*RnR!CUI(!hƎD˝)Z7M4=n]G9)5 {j@gn*ZP}\Goֹ5<7BZfA۵T@$}Aϵ5'vJH0ZPFp;p˷}q?_YǴ0sL23 B]#xBT%8C6rDh_L v (u\ ki~=k-);?[As^{+Ę<Y0gC3*3,>w=nHJ4𐗨# KP./h722tp[+Ȼ&%\+CPz^oJ+HσO\<:oߪa]l:'ѫXQW03)t!@.: `aט-:x4 XN6TJ4Pk7#.ԅ. kgV-* m4V_AFB#R [ R Rokqؠg!;dØ_М~w~yzi`z;E'Ivӆ}pkܮpo̯| r-. t{1N6zL~N VsȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 ]ws9q. LҜ@O D9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@O }Aᒜ@|rN |1N /@@9PsȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9Gb/ rbUgʜ @¨2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN sȜ@2'9 dN s='Яփ/\ZߟWkq}qiA_u1۫u_{ $和KP/ǸN.Ƹ>q (9q)>zjbjR ho ҳ]]'t9t຋ݹŸ"]Eb/<W.<ۅ6. eFW_ ]{N}'\??E=+ Xߟ}n(BV!f?˵[ۊWW֡?I+?'~o=go^zkh~Gٻ6neWmq29Iho&=8'KrmȒG걶%kֲ4p}ə3!&= 1Vo6nXih')3M>}.To<0~} Ia\~賓u}|8nz.@Їed7f 2B"1ӬE2yv..3:D9UWC |ۙ~"TE{Q&߰aI~z-˲$g%X-Ea 4ȬrN;ȏ2׈f5#mNyԪĦ>Op{{DŽ;ВI͖p="[-Cit,a[%Сgx003*$o-b+@+{.؝d#+5'  ]\FT,th5i:]JQҕ J)] <,ECWWFc]!ZmNW򖮎$B K3pcO^mJQ*+ +H,th;gwBtut)ׄEDWXCW\ ]!ZP;,_tuOƗr>X] IMu ;^Cx_j۝&irZm/1t֫gI'<2,O%$qhLOߟtum^xڎ#uzmMnMZO⃢/T>#!+Ҳ; U3/F~5뜇^zQf[ 99=B 0 ~&Ula<@q%Npmyy?J9*eOZwW {xϓũأ~?CG T\gE.)]{jgmE[-53w [v_ 9~*)hݷaxyfEۮ+ ]ۗ-4ԳͻtR>oZ|_e4/5*Ztߗ8u˘FDW;k}!;t(-]!]qg:5•<B6-]!] fO̬ vkhmBWV5~!PrB[:BCG^'oC`!t(iKWHWDSK ]!\uhu Q6<򖮞4 K ]pm,thrY}1ҕJۘ ]pIG5&-]!]Y~-tAϸ4~eP]{=Y# 8KGDW)b lZ݂jOWB= MCFӴMӖ>F܏mݮWUM]MZ͑2#gk\ IuRVƛ$qiE/_ .%?*J{ŕyqDfx*3*\ *H#2-,K^_0GK'@ |_S8*|3]E\㴬 P`:'mn[|3,Pxcp>|3ep>|u}\pc„xfwyXݫ§opn!lƕweUnUQR*MNgh&ASB})!g&MeWQ Eܚ>sĨTJFW<Aq)珒 iE<zc aqiZj,?;]WV:[_X|v|B>,~ '*Yx rNgbڅJ{5Vm( i?rq^ٴ7w1. -\^_l_I[v: 6x $LkppkLOm>t&c3^Y=C-c.ʧc΢P խ{d~R҈r;AuF:~|0m\:{.F5r^QxbIήA4u,u@@Ax W"h`beN.!(e֘^?&TS}0p0& ujѹ^ij4H\peBLgE*l0ӥ)MNey*,@:JFd6 JzN+,͢&eqqzg8QJA茲~U*"XN6|b`}̇KHyuFi!L:_~}ćPԅoz{vޗNnwK<&x0|aǣ>}3 Ԁ3A(tϵrMp e7 l(R;\ |-6a3\i\W!}witRrT}XΏ# r1\uiA$^}mQ^D:*pf xt6vݽ*oje\(6 a)$L֦Tv?tit|kkCq|vjmc߰(\|?(uV$ jßHpչ0|&Vb[krsM۪a۫vk5Uay pTp "juƕGok]6r[ %׵N1:KFb| w0|J3*Cx eQ,?,ow\|ջߞzW~Wo(3g/a z`L m/;0lO}U -F|j}z}6yKGx[ CC?u#|bt9Ū.gr'2H:|`_vúEnL!fThgu՘k=N5QO^,Oh.RzI)%& B0J]&ZAW #kr 0[?KK OOzͰ J9A U.3?=DS3jfSɘ!pe]˺4ŷȐ$c krH&&BJlڹh`V;Izԝu%YxylYH(Gi&kḦ`$baRS@6K"qroRyNSbD|dZR: d$Plbt`= ش,[nAC8L@mR48[7h, "S0뢗ԕJC[Qg43a`6 p i(2Gsz2/|m|<< E%sEq̶at4Pkb_U L6"`MFf}"2z@p.II2@bfR4[dMhhldV.HwuO%9<7 5.-<3^(0gy`aO_@dsU۷69]{F%ndW_ q#P`20s2Em) =C!nQ×گsao?>{m"=Ť^xr3/j.g_!9;t3{?m:Sm;v;ŋָEw]\^R9k RlZF.jyg37L|%Z9*10; 8 ph_J= VY +mJv9 b" g:mWRimCJdmHGu_cIFHGy&8NLdlZ:fϺsw;i}! AL7geUs9ep]I)MZ$r6iH[[+Ov]vHА[لd("HFkd4Z@^ Ќq|}r1B PPtP. DdzWŠ>ZMpMzŴ8O#jCO-ɡI _ΩlĂ]x5Mc*߯PuߔpJ.ZL(ÌVHVMBmbe΀fc6܎QʹZߜz^GٹDxLI3D QH@V2kW٠cwJDQ 6ճ (ht<r^(2 YG]J,6b"ngQN'jõeSɃd_KbL$t0 3 !LI( I *T>4;ʳmgfdTV!Gm-~6dDJDz-Q8d`J(V3Ox+G YX/DY j3aI3LjpT+fs⍴ #]48XbH2hʓd=B LldcS ~v<]DeMs/zv*l|Xn{b)o܀J:Cq6{Ó@dy-Uh a2㳁A=ک-˩hGv;w`1,rg 2S,6ȮHC]u6bGʪ6j6N[-^Cc-BIڦw:s:M$- tV#ה6θ7PAVz`'=A[1嘿X f==8*kvjm?@zȈB>ՇMT#v~i2 UإgCuȏ~# _?s>r2r *O֏O?9aĽ!*!q]äT-˜ ͒oa/`z܂;>,ݻ?-Q0ᯅWdS[ˡ mo˟V 9);j>~Z?c\GX7/X"gla|?ꓢb/Z3ⱏi|^#/魨Wzq4fjB7Y{6oѵD` ;;cbEg,ST#=ޫܫ TK`N&^;E&;BuQ"bdH D&oeJӣ5yaq^txlH\mZ}#dq51'xbil$u~/q s/C vc#qVU*R#Nj@hTR=K1~@Fo*Vp"~m0o`Tbʚwz!%./?hƱr-@ӽLRf]Il S8uno0t{zu}>DžXG(Zn}lRFjZSB8Y(}m݆*k6˺Q)EiF})gIKɑE9XkMDIj2P ڨZoL7jY̠oGin65RsY%fڜ_z9@΃Xd/dV~-Kr2M'?XmϩٌrAWuW_Gs+|} f.X%NF;8yAz7]q⟗%[E|wM]"Cr__;_ovKzoǼ_<~黐oȎt ue{Hxu a2[YW}x~ ޾ ż_frYLoxz­ti#]*Vfք zmt';>DUn^ᅫ_xEo<]i얳un{_]n_p{#χdz3y9m;+^n zK!\jG/Z)kMm0mhл6P56x?V4D_(}40.9q-_5JutF`ս?rCT 5)B,G]D4&"%rA-$ 5sqFicm( (o,BZB%xTT0EB. 65gX4]M'|9={WK!fΈ[uO!O 1ps"˫ټݡGQB4BDI e2"mR(}¹\K}.d]JKb@WPJ$`)@@EFJ+eӁh裙8OkZwܫկ9nc9łJn|!f*y<&\E=ج;|ݵ{yut3EBu2LC6J(>5BʽdiO(ORHY;Xbn*$)@*9Tl@ 36¦K>=\~z=nJi82Hg6eDPF!3L%3Prm@c Lh[XBYIV{uf+qvhy6K㌆`j&rI\_WdaSf_fij/tbrsq>q,1P!$E I+&QfF2csj8v{d1T:F±LbK )ŠTH|zyb8ߣ'`[͎}668Xn-`Jjt!<JʩT+svJdA/JJyA!1U2#CiE']<־TQGQ| LIuamml=IsNsT,bET~PbEܺ/$CJB= "]  y+ȳ~0*Xtm1OFK$(v&W@J̤:x#u8O[ě_ud8[ĜR,.vQvqf|v.d- $m*j4p5&P+Wιi`aO6CLXW % ُ/hxջYu]r| c*vJSr bea49Cq|Gh*Ž\o๼\3أ5kyVK;?l=jq[b3]QP&#" UT6I%PFg(21-zKʧϠ'ddځ¼ êyc/ig.ƶEV6Н "v5 Ŀ#hp1k*ܻTgۥeFP-h $$geQyk2 %PqS0жA[+6$dAgZ{F_!iޡn2ܙL~F?meJÙ$~IJlӖ%ږlmMuT*UёaSRAJ))$-֖8w( 67v֎5ݒnt8} Sl;DG`NVʠm~1 !?*gRbrWw$+:9:g c: $ ,28RKuBV.VIPN Mc0k/%%\2N,FYcظIPkM)B9,.< vlAË pw5Z&2r?^NHkDSDXQ׎FiupF1c1 )!R'wƔ:b(oޞIyݺ7뀆MjMHׁcy5+,vem~7OHEEjᏅct:=8k!bəɟ:7`E<<g(%br<-B0C} wmX[B9W&"2,J& `pr&AFD gQdӺ4JDfTyVvgl!9WMe.|d bT RZ63x6X$]i._P뀡RJ4 HyJ8B(D@<&e3S9;7*'bFgjr? l hrwm~,U9wn0=B" Іj=(O<>og qS;tы4PdzYT^.=C7;S7'.taP=0V`y1F{k,?~+\IB;},3eDv<}U]YEb%ϙ6+=ށiN) }ӛ^m?GmUo1Eo'[ɒ|#K7%["[Ud @0F@IV)=yt쾺.ty=)`/61il&?ĿƸC&=W[<;T }>Fikt9zj/uWfky'hܿfWn"5P}G.Š_Oy+&5f'{`r_߰ĽN+0KkI6=EMuwTin7ٱ`"ZzBT .#1,cX;d Ri\Rd"y 4]<5])شo[;U{vS__O6Q?&j0W^VGS\2l!sy1"Lhel`92%`̇T~ Տ'6V9JTkLb[+"9HHokEj' gN9=Fκmp|"7'jz `?oqd=>x7Na)LY/(K)"ʒU+1KZdlź fu@_Ђ{^;.&w}"p^c/<(lr`%kvq,TƅeuL9 W`A[f,£Wvˇ`d6ۉVj$VyNK-a zQJ&Rjy&ԳFp_}/^PmE HS#IC릫bsj_y[ b 5]*~UjVGޕu@7hV sq17^)suk6EF[zKkSH fcu8/DMYMfύk>L0}$-GWWBkJUX$+p\\R+O* ,UW \%i5u)5\@qg\񽁫$.VWIڇv]$%C\=bk=;D+pu?)zbO̮'-"bXW=D^ZיˡOz*⤗^zZz؟o:IGi_JאyxcccաRXٛ)1U^rgSx;cmlRYaT矙*B0yKQd_v+b2yW`_lEǸO8G~1{ֈ)2D6Z2>sm_׺L`ӛW*j19!!eu0B1[ ms 5\2)np09sH(G‰uXhEnK{ M /7IĮ$}i| D \%qU"L)E,zIvR7g#3i KҦW.Nկ {,Jys%pDrwx&"bZ;|U5P4 %^tMik4Wuzs}y)‘7fj. ΃SH#/441HU!GJID!{Lb{dzVA\Ծخ$-n/vaC7ZʛO>f\x%D̙2&{FM6Zd#Hސ8 w9ؔyS.]yTx4eosBݗr-JFqq*$eJX]9'H-5*/ڒili0oI;=;V:qIjGKo'SVyHE!V̠h$0LHdDKΑC*EFӖ5qԳR=5/2QSEGХJG~#6FIF*ڑNZU2k^x #oI@|"CΥEL|`)Ӡ0Xqb$;rA'u~Byw;5x=9q)$Ep4L'5)\0Ft pTJF)WWOVnVȁ6<1(1;"2bVjeS,lySqst+"mT`4)! dg5E6U>ݲZayf̂3I$ qR49e833թG{ cv6C6ցP:zI2 ҅/MɢX" QK"ҠD%:eLtz`"lfȔu6t:nY;D5CeG\.Kz͊ӡ}lXh[6y@Oۥg%@\VQR1>73@cnFXp4F~ TDCWiX]醳/Qކ,oQ4%sl%O:̲ճ!^5MC_͇~)Wz3~ee固[nl\&=B!)  f:oۂ.J{bK"Q=ʱĭSsNAS:6L+;}D(;N~b 9Ci{*rB{jlxǫmA;T F56btRW!dЌY] 5*RI4N['UWWc[TcKFE-n͡z+7=;jVM9LWXlt"EGQT$"KLfj }XF1B@;`!:[TQ 0b 5 e%͟IG C%ьK r28 ŠZLb.rj4sl:bdZ,mo`:4$Clgu194ƜE枲l8Y 톍ҿ>5G5>h- FH\AVIg _ Ȩ@s$R)8e=%6AP %au *n)&XLXb#c_,X- {^ffܐ7;YӶ+O>M|dß5l?ܥOJBΡ3kk[K'Y d jXQǐ=1aqS Mބ0Q|L%Ճ8 ᆉy2 Ma6-j vӈARI*sg5r[c3ż:W^a4ޣ $ME*BFfHZўGX B ѣQQ[ a1qamx 1 b'"Yk"ƒdEbc8GKa.l41EkBiIF@VL^˲qn(! /v4A 2θp'W4 oCrje2˼j|z42Krqs\(Ku6&AdcSF nu'v\ h\zBqIp>t`,;[2mk< 8ZeM* NxT"K[HU. `+ 攵!?LAINYl%P"}P&R)W'#qR( oZ8G&G/\CL$]Ԃ'r>(!2N(%ea|ByDA$\ ׄ)6f[093q\V۷N-;z v%H~4vzb WqZ\9R m8(H!Uyf)x&'Vwu -<Ak+ީBnð{er(K?>r7&d2bZ>r!_tJ9#EU%<^E B%"bKr^kT S>8a|$RFĐxZyY tԌ(8ECtA{% cxu鴤'@ | LGa3Ź*`;*AqfeM&KH,*3xYڂpI'3z/&9~&G|8г؆x 2| :BV"_9Re"4HCHٿax"AR&yd 'mgXa)y3PL̅}Ԍo])[$JWix9ky;;h$bu|/eJ7c5%/-{PgTëAulu~"< tә4m2ٿMS;9V=~cd~ Y;v].ZȍMV|ffs׹6ּ+l]n]Zo`epo|uز–ķݻx3Jto<rMg7z]NSVs1E䏼эްrh.nt׍3wXns}^Db=\zfKvr3ָlݍjԷXiaVJ_ F 2`]IW SYfE1aԵ{Oc9XzilE Ω>Da:XeyG7ST3JInķbl5z߫w`y_ nMVkXRU{_+ ҅MlPV@V8 DID#e8! 40'$;"МIT$ΔDBe%+g/YWrk{YP78%gz|+up`EɣVߵel R[-0 BQzlhhR>>R{M-FU UK\kv?6'Lk+EIw%@ K$oSB T&hIRHIB$#gp))\Zύ$)]N0IP⿐OQ hC:VTg;BO`k[xGh |7{`K}Pu|qR0ׯn65lw¬ؕt5u<;} Q;z7}=Ô3ٞᦲ8T"eގ1@k){Yx:ɛwsk5:1G|TϠi<9OT=P̜H0e9Z+ $4~F#BN*B?=!6>-ug}{#Ȉ2QŨer D*C@pN8͂V-I"Q![`h!]9DC.ǽM@ h@ @pL }"[GZG"z5'DWz5A&O?ݭl*AD9{~En(+Y +74QQJmsevVrlYOB ="SC%!J2)q d\KÀs.]Ҁ(Y*9Yf>}-ERw8y_ܪIBWq-4B:x-O9(LC=u8 [#t4Zj Z_L7(*h]G 6DJ[v=]E݅,ėJxVy}OƙWR 2n5bE-'RbMP1_ZU> Ӯ +pD=HN-8u( $(IHg}J>ko;ꬖAOF?v>=D!dx9R %𢚎G$ "唖|:*V7^__jJx`r9Ώp0Q_Gz9$@.%YzA)yΨ~7 o!kM?p%Ͽ. ~/>?qu|E:/^t~,e|FJX*( $Ey0o.X9 wNt럯?F ?kTX-*8ͶɆ;ݞEhȟKם֕ѩiB_A7L.4w ??tB^ 'nqa{SP]zɑe:Uz3zwK5[Շ?.hܻzS|"˛/jr79`"xzDV D%9{BBPmk ^gg Z1bJ !']T@k4/66Vdg/ﴙyc#|fCkRD@IԂhSE2{gM'YCIMRJRJ*ibg!͠lLҜr5Xٞe/Y3K=iHߋFyUw@ L 6::*%Zؐe1+hT_D{of,-I?V#PPY΋TiEqw﫺_+ ~59[Wo+2͵WY:\M ԒRVw=+a>ԩTϣR)"F:WxTE;PZr[I)ʥ!L&ńPH0ihٸ0Ѫ9;a}"$O4%N U0hC`4%PwI e>c !YNe]"h8Ҹ;-on@kѩϥbZX:\Ŭ7})1ޅaN+)UbIJODHzmg陶zTLEdlq2iP0c蔙%4cdX lMΖvgMAe3OJ<`UzGᄌ>pJFRT񏻢k@iUS-YH`p}'Q6²>R^Ѡ eȄebFj#hNfg)rȉݡ9v.O{oԅrj L:kƣ˔F)´P K0Q8IbTJ6C n 1S1$PN*^ +b`Z%9Ntֆb*+`e4FDr0GF[wXtTxeB-/&`QZ k)V &Wi¥ "%8NE#VM;bŝ&D 0IFgv$2i V,H;HpR^r} ^&+m.&m8ЕR⤲`R><T \uXT*"X3h'~M LOd4ˢ]RNLR5Om?_0p!}wq1~{ o|R\2x#p Ojv׼33b2Uj9Ѝ/SߖgFHVSY. k-8pr7~酓:8=A)J*7Ӽu0w"&#}N_ Wɴe_>Mxa&!$UÃނ"gV5˸`z^Q3BdJU#Joh^ɻ4rv t~n[nmd Q`޸4x)YUb+KWVb ĘÊq&_g =^;]U9#Vݥ*Uu%ђfc >f=-G S*_zqzg Ç߼{??:ߝz2s_|x f@ \^ H5`n?>hkwU47+q OP.ro>zAՂsk58Nj ~5oz|e끛-YI ljw:4 ߮J9p+ 1Cl.k;HXNF&IM`^Aə$4dтXoQQ)XY#ρ;aw뤧68(FϿhthx4GRRMUR"} 3 >3n_g':33hZ 2nmXTvnQ:םqݱqYYiG+wcͺT{Jd"3"POC i3ihynL;:ECXH5YL`Aq娃ߊXufanOL v2H +/5 OG*-хHI)C$ М$Fݱҷ38+}Ű ]#U&<w4 *ƁDWm!;`, t 9TA4R̞E a פS(M6Q@D9IhPGJ.]nk-uSDD԰&6)ϼa1OeFge),#[;Sx\u< i9l=pk s҄{>D0t1S܁\c c]o:rۍh[s+Ơ<@I5aӋO1Oj5|8Hg;upP|yP?A;70=H1G_4|~2Vp)y}RhgqNCɖNK2E8 R Px5Ϊ ึ$ 8IK>m#tUcw)x?u#ϛE-Š4\qÔ:.~ov:io0+`QEs9 5C럾/+'Cit~aX|xb4CʩnBT kBF-]j(]!J&:CҌjk[DWhW1U[ RNWRwttʴi1Y "ZiNWR鎮,S5tpl ]!Zx\UJ>!L?~˥T\Q9rzvxPm+̏Jȟcɏ /●_ejCL#Kʠ 9VUh%ۛכZݬt/M>aMRfѴ\eGw=$N.{:G >POChz[1 .1ާzQ/NXSN(T:\:xE1ܻO&aO^ 9`+م28hYJb>^髹r*gU}#ֺBo.wR֨WQrtBi;pUBL-+lIk jBW5%7]!]q<eZDW؊U5thj:]!J;CH&[DWXm+DkL QNGX&J Jl ]!\׮%jJ AH/^wEWWٶt()P~v*BF40]|J•]!ZeNWЎJ&5ew:&J=!aDKSTM3yHB3VZG?p>ϝet@كҎ<ooԀ-\7p~HM>RKR*Ht2PB"<+KHq@㎇4xulT֝RAH[t+9;#\|eNvLT//L2:-W2?jZ9Ϭ>vUj(eP{UnTe65tpi ]ZCe Q2J66.m+D\ Cٴ]= ]qi-+W슮*-thh:]!J.:C«"i]Uw[e+Dx QrҕT)"ԎpiklW7Ԏ(|";R#씮⭡+Y[ >W=+4'mҮhvpj ]!Zx4>ҕQh"~ rB6lWHWV+X n5%5ADl Q 7CW] B5tTD uLj!i5-Y<U/˚}w~]a;Xfd&Nn-pK7J9~Igt+DƂ2'e,p 9EX2s(qK3kR$T-Pk.|=Wgv=6CENߴ멖FD.m+DL Q6Z6T߂&\ڦ䤣=+NXrBOۙ^[oX\Vay.^}q#HaָRxkJ-L$)9F~?!wEe\UXH_HoX%VfEg Tc=斃u7Ó72fJltfKINt ! TD)2uZ$ DaV"͓`v AL}~ C堨) 9pεXZAVށW6EDKJqqƲey0N,__.*eҋAT*U2ȉԥ6/3G 1h3 YiNuVH?1{t_@,9~\q4IgH'ȏJ"=;| }?]w U,zT Sf3`^Tt !dygFrx|t Vdy9$hq+s9>;ĸUA΁Jk6*Fnʘ EE&1h-M!*{+8O'-!1͒j/NiU0K}f8(/uQ-rL܎/A IA@ox꫓EK.e?k_y^sLx~U,*MVt3Gzvt;EEjLi’`-^kK %i1h'vVL=_߁-b[/ [$ҠspBQb"ń3JS^ i֙@Q{/?=piv7G梧ZzH?V)(slg`H%DwT A'u~xgяdmegT3} QZh=úϥ:F"5 RD$tOh<"df3 ITEc aY ۄSkht>gŌ"p/Zczq `> fLӆuꜗAdN98>[Q\)AnǀU'b|TP㳅΍oEL(00vd уp"(LgcIƪ0=uӐ4)i~gAIX~#lR5My%]<.>Døf7˲ro$#;F#`p;C rv=׈\Ӥ^#lU/Lˢwg`GWhVE>#/Z.{5;%t0YYY{Q_l#&ބ#ozi 8f$zmJgvX1?腠? Dp&xd!̔}jL@:|ߎ#7<;wdd{;9f њ ѤNa3OP|De}zջiw}vZtH?Clҩ*EV !:2DEpsso6wG#,Jrk)6Ʒ֨zZf=}?\0=_Ի7vy:9!BÓz!7؇>^o}]emmո d~vlޡpۆ ^X<(lĪ}NZ'z){ҿW/??q~-<,'[Wk_$Ғ=.RWLUZۓٽ!r{%kEq7?vWيp[iL4Z!c2d2':wiϜvAG̟F?},m)a)!bITV~ $Ѡ!#+#ڦL L-jEA4%Չ".yg9*,&Ξv{2J~ZRS (JCAyHJX9˕*h,O8"3]/-_ BzB:d?qޙZpz8_fZMa1Aj+q;ϓy^k78mOoGcS g!+c"Q[p>+7m[)$8 Bͼ,7kR4By8ƷͯT~!AysTN< tk3m67mКzIωc^v|fl8ZG̬ hد[4|a9˕vy٫N{tԛn2!kϮW.Vrh!nQ\vkw{6hnu){l2Tغsjz{8ɎNmvg7m}Mز–\[vƻ;ouݣ畖a<noqu{q<~+s>uGdzmn AQoY-\bJnTmkO6 1L'5Fڢ\Ievr~S}˲Ac&rm !V0]"mjzH WHuUqO>#ܯ|i D#2V hL0X"q]$D*d`N"ǜ=Q&˵ VoSHʱ(hi-M&q͍vO`L,|3񳴚z,]cVyI9sqnЬ̞ϤyZ_AIBTh2 BwgS1q:/vhJʘcPw[IMT9-pj.Ҕ!TQmf F2ў=wOȕfB e]_~;*0-5* % S@FqH/)FD𮨡 2b - 2?KB$a%D3.1*/ךPPǦFUca4$ғlŶ/?W.xG9qT޽@L,PO!Jz"j+uxgQӐeﲵ)=ɛ11qz ELJi*(%XLXbq(Xh:,<*5Oo33n7g&a拼A?sF/ 96v.(NH\N ddDVg#2m3-UXQ'!;{.2½g(&GX ' }|U)@"_d+.0Rq(j¨m;m֛AT]itBp)Y`OuQ1j!kH(I*T̐ XfB8Zc@Z a1qva`R+*! Vx["<)K=2s+.?1ޓ-|̷\^,Ѷ*439z̙e>=~y}\+G˺\B] N0-:D<&<2bV:qbwCD2mp&H5ɾ{x2l# J. 5Ҳ,aJ/NچJi*K; B*% *FuKQ?39^#+[ɨ !*\I 4ϑC猪8G 9iۂN iZ@s/|4X =hs1 BBQ6#0/ 8eA*!KjW& a ',jB8ߔ;e*ku8vN;z%vp;% x1L9QXƃ$4ʼn^x3)RR\Ǡ[2 &~S|5[4_^J~!XLTkɘ,eDCGAOi:`C "XQ6L]| kO}W-O9Ddfe4Kx>4$90X23Ob=!{C%%J2)1 s--3s6ȊS's 4 \mwSy2]wa_A_+#ǛzXЭ7$cB4'dX rupVD)8j Y}1 wFbԢf%C&Cz)*na~_*ـ獾/Ta;q2Cb4YJHDht m% Da:vR%e.Rm g!"A@˔0ٵ̠Z1* ݻʋ +p ey?^/Q֝oko B ˿cK:*|utٳlѲ9 h =n4 i5>nd|6ÆS޽z~\|_͏+T}S͋y.Fz%J(=PDvQWf5{5Yh~o> b488-E3Ct"4-5`^Zxhv=ׯ9\O`2ѨnnsEjL{PsRm1Y=Lf~҃nnqnth#Lv 4On9"}}/WuёDmJV-71@03L{cyJ*㍧Q>Z_:/QժV~BBPmk ^`b Z1bJCKL%[&cCY#u.\{´ ;o1%G)*](UJVrg,'C;h— Z(ZʀuQ$q^ .L3(!}/ZUN-3eJx9‡Ȣ3LVSB$=x&V))Cugzwx^׍_# ]D 2`=PM9٠8ꭖj"0p,0RGDuZ:lPPGpoUu갟l7S-zG`IyVYi\АH~l[ Z\^2V-qe@:mN[݇Tmfw}k*, d|O 2$DMB"XL"11 V2ΥHeFiUI$3)J}H?JFss=l#إoVg;v2wWrwD~$C;>JM}^+LݵbGJ *Z^:l(aF6"'&͈uqýFe)ƂHBPB."DdnV;!-yڴnZ)V<޵9&v~mxt p y󤝉ނNk , {j3}WVAodK7ՂІKn/^]6 |7wh%ڽT\{xn1ΙY*T`KK D$2e.c g@t'eg'Y6;w6;ykDFU. Qza<+K-VdMGh:X^V'Ku^6Նkㄡ>PB$0%YisEˀ5q|ăx8_Ȁմ+,cO cCRv{uMI|խ9\2(B4Ș#A;gE `w6JRDPwx.EQs uhjmcc.Lp \F"  e]iMo̸tISkqfhnBsh9>*e$1AzepHɥ 麅PV[rf9b3oJ SHSd0zHxH]&9c)gYg -)w֫j`/Ic+i)l=s)jh$ 0bER)n(uC:6>Ǯzws=/` Mh+pA ^0OC jjg4:ӊrb+j/<~>{R%Ϟ#/"xe?59O?a-q?q5.wKJFχ,Q/*TZ)\h+=&5qVgeOHvy~.h+ۈ`k<C^=x|s+X~CշՌ͏3? e<[Lb:(N^*|ZO ^&#<8Ş8/8DyhE{z%?Ŭ Nn:ҪEhyBNTȵvc]Z[wW-LV-$ΪN Jӆ3bBYC(-Rh.]I IF:,*4Oi΀j$uI0ÿVg;肗Ud|[ ;q?Hs?^?shHIW6y-H.ZR\Xd$FKQzfe4LXٗOS>9˶ nCϊσb1&f8c)d<@L~5^hN?@ؼmt_nŠm(jPO `Zt)\Ԃ_=`!u6Y WMcޠkNja Æ㨘uy^? !Wv{y~wni4</WouiF$1㚯3'xby卽C^ÖW0m5.E]Sua]ھcRI´֭C0UVa[qZLܔy٤J%s?50f;1D"E :Ro:To:Ac@X+y1r Tzr)%](2 k-U=^>2嚅~ u9eYӋnF>2P{,Vú_b&1+`6~Ѧ忦Ji^PJc)JU 8@8tgvn+ءp4JHT Xo+FkDAh2(($MR2vILJ!H Fy7F%ωO.+dp*#.qfk\3MB7Zxz{2DOy^۶ɾmѤm@k6,X0u bzzGxASrYRч ZhH” ):D}܏NvQ}Xf"DKK 54V2DS\"DN}_{E+kڧNJ[ox- tnSB&$XA@d y!zy1\Qt-I8?勲2&<αY[ס*PB%ޫē]%0Kٙ93:9(CX&VSllF(Mٜ[E87ZOrW+A˕ >?yϖg{!@Cmb^Ϟ|3AȦؒWtm66sqAE q4l>L> ,`s 6Vꦾ~Vg9:8|9| W*r$)=Ÿe𦌓-7U8&7ӅL>pޜ8yE_^}y7y=:'0>goAEy/!WMɏ޴1m5- oSn7i] qjnŏ?Y].V=bYļ bG˯بg5%`VT,PJg2!b7 E^Q蚏GcD;0(h.RARJ XQS9тGwO䕕ŠC,JEC5UQ p>j$x}q@'P^Yp :|1 Ckck;20";b^U}ݤ#,h:Mgi:Ѵ ٦'~D G 6gqҲg)#4W=2W(ޘ,.\eiEURx?٫vܠc/sk6D9*`5O?5Kmiֵ~`gO (s%c#IPZp@Oн^6ֿ^jz2E2R\+S:o91ɜ{X#E=Os=s nwHe=ϵs=s=1ۃfJ=9{Nӿ9{Nӿ(^(m9S{TInV h4d+S^Ry4(R( i-.lF?8'N168?8pN?8pN?8uzoϋ;sd^-׫lA텫EE{ūG 4 O Ϸ|ީsF$6OΕ4yb(C,o]{UkI Q{9QD #PSdF NoPJtQB6.BbIN+RFCk)j*Z &yY56S'.FwQ-u t}uIFei0oOYgc*^yۀ Z@=oUdBJtL!@EQ|N:!)i 59kRh1KZXBqq*$4TjzuB>^b|LjB&G*sQ` ֔*.c;r{ը<Y.D]%U|C)hiTԗqgMD}ε%KI)͢PJj]O($ ,A! m9/U81htx ^'N#ZRi(!eu5}.Z81hP9"cCMr3_$oNlZ;I -/Tzf$掔!<8iгseKHh]HdfjP.608 NqY1cW=K&mZ5MH1TNAF4٪u[Ύzvbdٺ A˧Q&.hul{u98Jk98 !* _ދ vov)&*/:E>9[0@ juZts5'$;[5R4jJjm)U ʜ))65U0] Iqtw鉵S;ne>T7ɑaIҒm 4󀋳7>ϖnw%U~SYv/ov=exngYcBŹ$1X\aZCm-[Ky]>= ]*K+oxvv#;*>'rv!/mn}۫}7=cvuxo"*q=n_~v[.#J[47XmI~U?)g]Kܲv}zt9N[Qق޾;ԳQ"P+|~ϔ-pf+XL}l+ߠD)J0ڹy"8aM5vYVVmN"6}] sI&msTEc)J6UC*DS9D1P]Y2j@A%;ms[nͱ6Uj:՜&lylz7|v[Ύ Oڹ_)^O|4Ń^{7㻈6t3(ӟ>v1".8ĹHaӬQ(JP9q;%SM;Sрľn?%1$+tY$-!cT@rv؁h(Ls}nr>}!V|bv3M~A)]-cXK9Zm0,@T ju)r :ؤwm-g"nX-&Bl aƒ»{ܐRۋ4ʋo Zltuurq-DI9"QP)D*JQIJc+" *cK%B_ :y -K(ʼnٴ #!oNVJ\El-ygAy2tka}J#fΒMds%L!GC&:R|qI qՇHBE$BdDd9`b䃔Km$ Y=출2oDk0E&{ZDӹШaE|qUI6&1r唜.DL5~*(Ȗ[o u'zg5-љL.qࠣe HڸhS< :[nف˫bJN)MJ].$T X]2H@+7>)1 X$}*0M!0jCSݤc_{h:C=< ;zK \ȝO~d{Y VCޖQ, w(ՏG~tqnVbSIz6kG6y2yr˔ D⁊rԣ0er^*-u|Dӧ_ s̟G8MbpՐ *3j$g8e ERRUmEe1Ԛc qO(z1(c r.$+'VȖ&O(`ѧűHpIdo$?8wHΑSv0G a W}LX#MgԈ}$Q~ iyfD4x]dwֿ[ BJy*AhG$?˵Jh.&⾑K:d|P!+t3VB^֦ca2M;`YMKh3 Y%g31_Vp}~ Vo~϶߉C`9tt$tT0`\' U9pQ9:Ơ=?Xy U #؎Md0ȔX!r4J"A#rƠ|vVjE㔬'bMEv(kTX *j,X>E_:*~{t)NY]Ζ\݋9TМ+l9Pz<_BAx'W M٘:a JJ-igB?^yWb}l[?~0걬R b!`4#ma~[B(}oןo7!ݴ$:[_e} "%$q^rz z]mV_ Y%N(L>xyBhOeEt}dU{*(/"yۚ"%Wgv_1"\ZTwy`)۬RPm==觧9[<3ݘI{i.'\7>S??yW|F>-'[_}7wSjg&^UĴ//ހ-]?qȝho=tM6Oޅ}]zW,<ȴ)[KaL[uWWش4fr)9LSF &aodęc-V)"s+ՐLGb# g*N!цrb1xO/Ex˛}lf7=0ʔ>|ǿ%oʓuB1NM 96똒 b!'ֳj_3䄁Q;%$u@*'1MTTony^7w-|ێۼ׈7bDG,Kt馩bJW\x>qT>90;'U,*;kd[5ŷn o@ DgXh R/xTg<]5(CASTXYL\T>Nڶ杧gFBݻc6n&l]`3O[YR' _5)ْ-; &dUUu9 R(\$JKFy:8M+ WZFfu C'0p_}œ`} h(G /r$THCKK)FDO5Uԏ[E1q=-I=%I$Og1̬Ǒ2X* i R"΃4!"}a4^JJd 1Y %ǰqAP6qփ 6KQ`S[a+|#c>V2C qHDSDXQ׎FiupF1c1 )!R'wƔ:bYwޞ m%^6YjP\u|7Hįn''3Ojh  ]oIID5տ0$jG65X 봟xW?JZnVo'%&>y^T DA]JzVMs\z_+UG?L)8P4# g" 4w5՝|-?L uq<m\WxU>΋ə5=[Ԍph<(^E=&Q]iʆӀg u=TH]L|vޟ D~bRF/>fa1Fʔ~4zˀ<>bmC O,ʢdÄ8RcLl#f(M.28Ii<{q?~]0׋vs9Un,һ¦OeÅ\9Xl1T*ąD)o@Úl%{XX4o71 gKuҀ':LyBq$1e0 ^?)P+~'*(SOD4)~%(e57J1x1JWI>d/[?.⟽ֹi'm?l}Wjzc[ojC_̞]&ꅕߊ4%Lr5'o_>q0^J3^iILʋn8 `8 ѻ[B5S ?MAosNz0j>quIpݟC!;m}!Çح~u<2K/okc;ӟC0U(N/*}ZlKk@wQ*^X󔆶/z痭W[j~,PlLٓI۷(}or}2f}HlD|skBm?/7߃7GxuftKMX zImD͠Ly'wL8fX izh;[ynI&`r߀lq]]pzلYhE^1ːNV2sQak&%~=ܳDu2VeXq .= 0q kA :X q X:o#"2M7K#Ӳmin{s1aFGgknﴍD q۪'L{QJ V>S "`ʓ8cn(%-cɅP̄VVHi(%Mre۰oEda! ֘$VDb)s>`5-N;)Yv`p+#i){۰-a[Udv#빭 SXb C z+R$j8JG-j81e֨N@('(wpϪ&/W~j]w1R<.$O 0IsɆ ;XҙP[w>;k{70ð)vv|V}xl:W=لv=lzv˒-_ ˞mWe냟_^Omת1]r?<ވuq- 7Y :[ub7o>x~WP[m.+&f`^ŭ^8xiFoݓ߶9X]Q%cYW\I}cl[u[E;{s%׆FI2ySA*ͺ¨M& Pe['ꃚ ϒ>ݸ؂~,{1 '>u!?}J}e M寿G/w=J恒+\͕bU+* Ռ6 z!^gL )&ͪUt;/kI;^D5,fHf-Λ5g+f6$#Z{ݼc. ln6q6. 5e>~I"~Y 0Uܶ{bh:4Oͺ9Vs<.Eqo !㙈NhO=Oهh./% i)Kge 4Ƀ8䈗Q ($"wb AOQVa@X1c2b=6M a, DPB$NFoGÕ,wXnkoZ]m]PW(aI*b< n:A$Y#(MN! |Q`ÉӀC00^2]i a5BTLK48ʹcl)aPLśpyP}z:sB|P}}>P)np™9->Db]嶴u@vt5r `*F{*E%L#A BNaJ9d<3II8 x4Q0spX\ʂU)w~EL2gBciNupj$Zϝk~jҢqg(枋X3ʿKגrh5m֬H u1yl9GuYH|޼Eogo{oqgZ^ogeMc A?q=͔YnS`Мꛯ[Bn?P@6!@\0)6i"miDڌ;;]3'˧?NuMׄ,P2mi6W)R3j"qM$nOvW#D)ʤR5A3futQ#kTR% A`n1:gNT *H$+y}더fMΗf<:мIt~d{޿Xlt“&.:Aǐ )1KcP\# a7ịUh4BkAfT F刊`) NPJ25885x!zză>dWy}ڕ\W4#yBva|ИNz+Y(OģX2Y 8XwMPwwl#CH 1D(O2Jq]"mrH Bb K%AH0PMjS;V//1ğu5TzvKXjS8 Qq)A nA $kk䭃dHjXD2 J" T)J]KNU0ѰT}efv/ )2x?Z x?y)ӎ=HpKypX>^b&2 %ZI@H1\Zb>jTMȌ ^`A ƜF*e&z˘dF}ml Uc wUh䆩MC7Y5Yut:N{8z$R9Tt'(rF"1i)BPd"&mSdTȠgZ ꌜ q_k͖y{lAmA98_IɻYVd&|DSDXQ׎FiupF1c1 )!R'wUtC3o'ho_Ii5\oe,iK޾y&O\˕JRi cLάN չ)ĠW{{fK)Ďx;!p< ]v[toFG!rJDeXY0L -`2M#EFNJ(풪,n*1(f 6gUŐ}eevN62Xl1T-wRZ`63X6X$i-_PK*Ҁ#)u&jRMP>Oxxzгz!`F# gQ)mcS5QRV f L}=t n^Ϣ|adZ/$S؅sc̱|: :4^CٯSCM6 hM-$ s7vkǦciTLac9T|! TFDVYZPS6MQ MyfUt_M/~}eo&V]}̌ޅ븶O5ٌjK??=-JS}xc8;?6l~ʒHh{9w>r$ᨊZ$17'y(&7GT3kg sXa9MFezP$zi1:Gt膃40\DQ, 6;/FKb#XX>c0ËyY=$|x[J?fQWRU2WufϞ=>*HL93b[p@ofj}`x} Ú[?- ?+X}]͏)zK߱,+Ɇlt{9;Syn%JItHI)Ւ),WA|?+Ӷz 0~^Iy筭Wrx9`-?++| xx+@W ͠`05m_LI,6Jʘ33-< OExe=VI:Ѩ3*E ؇^/E?=էZml\س 31 ܜkl1(_M F'V*o^[y,޸yq d؋n6ڌ Dnat!(^5W.|3]ݬy@33,hD.E1!$T K ߧތk4͵L93F0R\mПD>֓@ٸx(cEqCrrMmF[amn&-}1y5_פhwm͘1zBrņZm3`n DD콯Ń~Y\Ll4@5S1-TYw:x嵒ZD:P`Jv@;:XbD; ?x0(ß_xN 8C |m/A&  iZ"h t5+Ua:[;z;UC)%7C\ynu&>G=5GLfGvT?V'ugO^ffa* s8/D lhnRzF-ۥMWp7*B0()% ]Y%ș yJ_x]<.TiBR^@ALpo6MDo[ O$ϧv5 T ]>W?{U_c;7w^W% Zil6.'5aOl\Ed`/ETw:=el"YϓOwx{KQ棁)C.B0rDU}r)e:ŕ3T}>Ͻډ>Dr;d}(ZOw5D1Xʜ$XM7BK"‚%*p rEr{"LG=۷_bYX=Dy #빭sgj ]k;ewح6_l_Vl]mjÌ$W8X~\L4QLXI&*( snSudDn]9aՑVxڝ3Qo5cwDs1$he3`!\<:!XE&7FMP j:=V)n[yڧTA1h ha[0a`Is";n;#&َ*\ hgV5\xxۛg|::r!Ayߊl^bEvY(,*1{}TkTw OQ"ӹNH *SPJwKQ;*ݭhYo_tnƜiQ3!wHD*XDփu1pl,zD.TDp)q:wI4#ZFk rzqyֱ97(s\>|p9:ƚX@f-tRh]k^r5y7c+Ӂ{%" olyÔg*A(/B2!Kb|aÅsn_VTxGDJN:H'T4;KI%pD0AP8] K> n(CQ vYoL[̬PL$-@Q-pJ΂ N%\8pj @qdO.RjtFF"B] !RLJi\䉱j2*Tز=#uH9b"knAvZϣajm- إW Q;n|z?JdЫυo~23Ε?7a]e8bk0ɗ<aU²qaIobƝF&#·{dmBչҝg٥j$ sBG@cc 9eSDz#;RL]L&.z)Y/ tk+aH4Z!jr|N )QkJ*p>;)[𙩒dUuhOE6243߿ yOR\p"1܍ZmMʈ?85Yq"$d"LfKԴɽ+7կՍIs`Q9[A p9sf.8(O깥rn$:C[y|pbB!n$A֎ CSjZ9byC90jXXJ|X.ډ\YMQ qWʠ-\V DiEPL\HhViRj?!m,s l~V?)&kUbN;(!ntꪍFZ&N;[BM~E"RV1$`iŝ""v8Wa; k#ױ鄒( 0X>;vbqA28Gr{+vʸ4qUuK"YZ2фXri1t/E"ϽRfc6Tђ,uvl:w>}c{۝ڹ1(U;wzTlUҝ)q(}ҝ4fΛ3tEjsD;^#;W;Tb(؉ |7OV\P(ms3n"|{ [2^lyEm )s25apױsf78MxuzrtX>Kv3)XBW߬T=|@UDyQ0:u툭EHi·S]-%5j Cu֕[NBҥ2YL Z>9}MUPkt-#'mǺw8D!jOU;gȣR>ڮJ" ظdclp^-bJ$=5z1$Dk[Om|=B=ܝh S"5":D[WjjrW5$Z1iy7YĔQXc ӝW#UJ@fK dԄS0ʹrt# $l ؘcr6WBk>MR瓏԰|pb,"fr nstV\2ɛ=}ËBr+ӓ;p6ٵ:o$Aw3,~&=O=Sb3LJ=wrxdQ76`GuE[~/vK"rEO7]}/~=<^::=;k{7_ل s<çq-^ QpǪ)ĹJ(B"[j .,FKAa4K$Q_7.nsWiӻ덱Q]znG'_~#࿶%x}pp%wԯ?;>x Nz!yTL2- f2ApPR$9 3SmaNM'V&~*pI*Rj%.6vΚcExl+TGdz͗ly;Dͦ"54]du\1͆[Y8j FXS<"(UJ\(#eo[Obiȣvs[}gIXɇۿ,,֗| ݗDbԝl &Tl |%b@󓣶Yܗouhqztx!>3>-»*S|o,^͢"z K5brWH/a8 GKUC UfEr\lwwWD&xD~׍Mj7lj'̓?)q} ]YTy'6x|1\">zWNꘁa6ZU)W?/;ޟ}z)}u>} l$/?iZuݰInthd$w,̥--q;ޮݫM[p:6 ]_z N_m8 _ s b`ϫCOt/>r77L ˳ Oc˳^c(LJ#{]߼>6|:<|M=z$~1p5čՐV®Ր2=\}pbL%Ur zb*+H$оHu/CK5lb㉁DIϐl_PR7Ag2=_n*TM/WOÝf0B`?7RѮ.q o+j94qw;ɉv΍^2驔XocJ)׵.s.W}zUԶ0xG'rPTak)??0#ewSqӚ_V/Vy6==fw=ME,]:g"Mل88dL QQyl&\հ.8Ok,α~ݑ;A{F#>QG{=]iiAh1x'%NZ);q*9b'1eiHḬb"9تE[5F %@'Ւ)Doaޚ$saIwВ15䋊-b;ńB:%mBhmL~Naؖb'/WULNQdӵXS1Uu@s " Z'%P #kt b ڮ 6hjXqvl\@V+ +Cb r-G`{`Qљn<1.w0.tvi!r"gUeܪ˃CC @0ٙı*TɥZEMX⒲Ίqah<;z\ȶWgЀˮI Bp’pJl7L6Tm)cU[b⩜B֮)a$qNa; Yŗ#נj&.m 8 m4DeSS9o87/w-U0!ЮT | dST q+gq3FN[/@fNahUX,p$j`&Ij )8C$;,\CTD 4-W0y=Xgr"\VZ!hJA̞.p3SQ T!# Z 4}!3Cet2\q$+!ݶ Abe$DQr6K^b*18߈U8HeP;1 }?'0-6GdXAwX=,?z2(ܦKj0moP\XYGVR%h]JQe ,Cy$8-xVaBgEլ8\d§UV!:l5(tiZXyq`{Q A1|*(uNn][@q;',8+j`"%*ScPۆhˆZa( Aw*4}ZX@jٝOFbkm+xFSh4Qs`\5Aٻ6%WX~^l/I2KR==|%JGUIYLOtwթ:0T&NI/e}ttQV]R"d8e@!J q Mªڜ׃bF`4F,֨4||hJe#s4إ/,TI8KJigD |w72Zdp% ~<9ye}p̳Xe0!pzw",U@k6֜OrEH}xa,F%jgB޼j6PUmeVae+,W>Lztk#I'y*C06t.Y7u0q 98[L!tiImx:)kg @][ 覫2m&xis`*@¿mŨ׺െ$^kpW  o;fc\ y_ 0 |paQ,T@s1ÚJ -65h+]r).,PPpylU]1J$a5>N( j=-jx}Z 릍M eBV#bʅs ~rpuַ0yM ;bۡE >RŢ45Xzk^AcPx ӏA!s6,ѳd).XZ#eʍLGjmRYH>9.J3)%&}*@$S5%i\4ԵoH;?l \!D mMg]KrG05H+wi`Zg'2YĔ9UµJIF6oltYVc Y)~u6D]`uW֡[&*3!K@ fŨ+kpXF:lb`eNW-Aez翰"m#BgUK}ljcY(qǂNJș%5JsO,el:_|G^ƽ¼ݡ_r$xgt%J͙iD*2%[Eg1k|[FhK c^TFMzcQfǓϛS˓%V՞~u518M_UO9Hn{bK0Acsu2:a+fֻ[lIeS v; l: ,Cx;J];[t\97:[I x|$yK.(836+Gcc]^\V`3Ap :6 OxrsC/VlYln}|7//GMFZdt\~GhYT>|td@V'O6ٰYMG2s:*UҾ{ߙdZT;ߤ+\N%滨hUΔ)8ܑ9nͮ80soNP??ޣMIovWtމfuZ7y;[򂞶KMs|Av $[?۟ГeV*g t\n7Zzשּׁy/܇+PK{d}>?j^^2 6xV&k#lB.=Nz=]/}socdqx5q;n[{ 0^\ wb} ۤ+ur~ d~:>wc i-{i7p3wrAoYn =3J*Š(?贋7>uZ] 6,] uJ1]4Jv/|^xxI?AJ=~tn`0Lcx`оgYj305UcE+6Gi"47 F( mJ8-hY-_4KdUSۏViZĪP*!]Wwp`Nv;0iq$tDݚCC7=PW:6w}l~C4 + g}aaTŤv uz*=GTFZ?7M\~4U=7MJMeo^ds|}}jLڹ;TsBV2UeG{eGTX Rϥe~@T%9Y6^m+P.9:XVgLfaFs;Je`K ]"(.cN!SU,go+5Pc2wx`0qo'B:Zͥ1fx;z䱪EH&٢DJq鷘b|ʷa씱/q̦[[{{χy(-[[BnNo!lK%:ib{̽͟6%9ϖ]g`6n6hc𑶐uWk+ "CVtCKN;aH3)Q`,"q"m-$c╊jUiAb\U/2h]R*]UK^*a*ވD N23*GbHl画k }r1X q}gN]ٛUr.gkvpl_|/Bш6L N>.q̙=,T 0F;̺}g%dp%; =e2ihI""ļbJSPE 4Tqw׎+ lS@?j8tl{ef * %Gq03bNZdH<`!:[6lmÈQ*Ԁd| 8~4B$a%D3.9&Kmg䋧I.qxۣ ';`lk;/hwH)qwז?0T޼FXH @t܁ "h#u*:cASGRBrd $NBOyMdPT5HIX[J A42g32Uaa+Ba, ~l8EbzsKצ7&;ZnS3?sF6$o5Z$6X`< hՙL²AM+x wpɱN!N0 ]>F QDR7bg3b~/ܠc)ڦEւ]c$ RR=;yoUN39FqeF=r^ (&H@!Ȣ=+x#b^.$F:bl _bCAbcGDdU`Eמ ^$F!:3NFSV-T|ҨhײJ'JIJS`) ΃**yI3h8YWGCZ,6JvEZi-.A) paƃ)$6Gzn/̧/zoEB& 9[XZ8b'CY:%%1%Pm$LdcSF nu'v\ H. x 8^$O|8ڿLLfARG0ٓ`..K֢ۓ%Vxjka\Mpv´4>Dz:N?uS`y8$(x9'h*KIô)D&i 'd|RRƹIR|PIs sgZi+-VJVmz+i#-I7>{r6NΨ[> #>i{3}^˥W%N' \qfkE )Z#=qOqbЌX'7k=,Qb`,$.!"g4(A941@C>19ZHΔ&(ZdR9C{Z_*&f#1q=Dif{ vlAV󭔼>>hCf>|73pe4<yDoAXIIiǽ[sCLQz{&'m7O3XvȆ}-<Ybvط-y$~#I%sM SRd>Mmq# D13{Jɥ<dz:.3nwp+S6yNs(I"H0e92ZF "( %&ĵ9z2)v3gtUVn}v$[MOEgeĈ@&YHe()G,Ȣ]D%p$}{AZ)D%C\8&T>ʈZBZ EyJ/Gh-6,CfiO@sD 2@VT5oh:O{WN" Z)<,гzZynȣ5DI&E0z?0K4U,ZᬍB3}y O6׭qz3ek=2`! ѧyBKC=u8 [#t4Zj Z_L[_ (a8dchPw2J3,`,d/Mo?3qPV ?$&XG$EID!wTh.' h`|IB$e$vL_ckwoSEiaPS%g vIJTo A7{}=di\1ً~;s|\zQ%RLh 9#T5u:k_;k~}o_T_Tݾ ]Snj%ϫ6/};ԣNWwN'/7nToJy3ƣK_Uxk{wgI*RōU'S- &X V@\3ǩkdTf&!Nw9znx;)O&;=|!5 NMZ[k_z(vӬx}9=tt;iX9siW97g'eLByLO߾LrKZjNnռ :IZZhkHV8"[[?/XV ^P㉵P5 * $hňM4*HL%$Ke@/@NQ_ 圀bccdF].8kI24ٝɚKB gI͖3I,c/ y<8d /F  ѩVSh]h4FP٪?MV^~WN%gn> q"o Ja=a\0=axsN!qx+ }R-͙;{ BVʖ\ffK9Xޡ"O0?3y׽uyg@lrUK^:Kq#aoU_RW_#Z{V8YS1봲NozǿӏwՏS߼~Op 8zV ͽEvi}Ɣjۛ5AlnSɧn:|v9-+]>'sD8?~U/w5az͖YY$X?u/Ta|–NoHR \*&hCf d1rEkN<Lς' EJ6HSb, ()C$Ɓ *(Ȁ.yO'y˓zayD{>ÁXj|*I`rq(,8z}":$SlL$y y}ws<7:e;ϭζ7#9eR]M@3F`*3I,A{fIj6wwgrgg<ҝeDI,yf[i1O$xM 5CfK,gkj,y* GYǁG96Oֶ$;Aa$>=>DaJi1\G.JMF:S+ ơڪ!R2Mߍbn1akMv4510Z|Jh'+jZ=NtR@80owQY̟9 FY:׃.86'?{~LUK5vuaDOշ7bys=]xL}{-';vvN_u1L~˦{V|7p3/wΖm3'_&gbܤ:} y/kn,wm6ل%fy Oӫ\pNҖ-!Iz>GX9Fݠ KVrpw9zh^_X6Q껮fhRDhz"XwD"lӍn2q׿l\7lJ+zw]:} c%dx&vܶ7 LmyJэ[ߥ4Czyf!;!LQ`}->mh]`7,n\\_k9/e n6s vl~ᅰ4O`߸Gkw78r#0 ʶ@Iu+!B!U ~aBC(\H5p}y?O<6i !bC0JMԛPv uԛB-qlܢ8zK6ENJ8 G&1Lxܧ@:}jL]bJ0ķ(mgu1Dcw9dl=C^$wpF=wm:r.Fɧ68!2!&oYٸvۮ?f6sfhǡ11Hm3(LǙŵ|7Tqf*"Δh01Qۧ59&oM'GN=9KN)93w4R!aKm7X40yY ]o(ІHCqp>gz{|M|1p*Yp:<>w:]\_1B4{?y15^](^]j5 AMR&O] xxSrSVXKt%VpZ%(]WBj(ҕoЕƠEWBԱbJsV]GWGJAOt%&KוPRZR2jҕ+] .:-P|t%~UW_8Rk95N`OnO曝G cms Пo\_gAY\_i}٤.ߞ 3W,µX>MF|+ %^' ̏}Ω u|Nz$.55f0৐!'$zb;f;( a0y9^tbzr0n/N#DkHC>}'uqN ,T {p;syޕ5Kfw5}ۄ~Yz0:FWKA6+]WB\ uRtV7ѕ&[TuB]14銁fZtŴLbdktF]ϻ`P]1W4\FWBGWBjtF]{R+QOtŸ]1-]KS*).Jp)j:SdpLUW+JIO]pi)~!P]QWɚh4؃] .x-'B]=ޞ|\^WYº /dȣegRb,KW6CWC϶2Tju&m;ȟ9%g"yꨌdwv4zb_Yq_O?G㽯5(yD㹻}4ȈPi ~d} 4|u )l[8R; =/._BazBaQK(, mrI+ H!sw,] .z-KוPƪ5'»#ӣ+jt%'2ت IW K'څvˤjBCEVOJpǸB t] %j -%Eb`g] .ѕSf]QW"] 0] nRdidP(-V]PW*UWϚ^}LJEW3jJUI3.5vuOgʺѕsݩqw_נśMg&tj(Xpއ<`Z:eG vyB1aiW-JїN*v6.۳F,r>YYlm'6}Y@M'KoEKmGim V6FWFJ( k\u&V+ 6&jt%Jh+PuB]!82t%^Ot%&w%LY5*iGoJpAMJh1+,8Yt!x).Af\5 ÄrMztE@,y&ɠІsWBkjJHղ{NztŸɨ] +>)kѕsq;sܙ3G E8nHܟ%븜Eg{TNcֺR]OAHugpZ/AfDc]x\s˄y  ?v-:HW P3q,] 3Zt%ޖV+П%x]p|բ+Pr'Zxo|T+VL6+;U]GWI+?kՌKZt%>+Z$FHWLzt5&-ڔJSzVjG=F9Z(Uv a@ѤWҁ^]۷ߞlUz/NW3͋W£lĆznϯ?~z*x pq hBz+, :pCo(ҕQJhC*]WBc u< 98]1n2^+tPuB]"yEb`rQף] -+D_uB]AdPѕl{k] u%T]PW銁jt%&KוPRu4Jaz,] nrZtŴ`J['kU)ҕG=\RjT|)5(}bX`5ܥRALjJZt%ѕFEWBKP2+s˰s?-9HUMgN-2sj``}pʟ` GaANٙ>X.J޸>ƥW.K--+23Bac&JU+P\ uR:^= 75ZW|t%jl@E`Eb\yK QXuB]AJj2(JpJhCV`LTj`D5ܠFWBcJª*&] pdPp >J(m֨+2!X&8D5\=ZJJ(S]QWɐRh!ÌJh>#r/vՕv]=>kwul2kU_HWyPBBգEp>XmL::s4:O *8lӭ&Rav{EMNd.n9i2Pn i- [S+j^ ^K+KGQ1u*ϣXcccN*븂E{Vlf_N*kc0uE[AQH(”QKHȴ`\!PZWC{"] 0 h[GYZgճ{+?GÓcJpz[fʚp[" |4Jp!hѕ"+ jB,j 0鉮7:SLWu]"]10:P+jRB GWBUWkUO(ҕ{FW Vv5y>\UM8, wyڅL u BLthJp'BJ(9tjIϒ66A< wՈRxb*㸁:\i1C{Imkҋnݑ3ɘZbcPR˗d#a~dK Y~}M2) [Ga/հٻ6r$W{l|h`p,MbYĻӎg'V%bu"Yv6/x2-P l#=?1O)MLv~ِ+Awi+usµŘlhu| \޺,gH ׯlx9ޝǯ+=>]9bW1iƌ+FyEUMioF'AJ E(PX͠%/EK(@%씒PX(k:[E3>ZLGygDX/u' Y$%Il IWdTfd9& `tLFou^2{,rzO KIJl%TsdXl8d8x?n4/M:tU$ϥ^-5UKGjh<՝&kzc+1_-Cq bєxx~i<|4ai6ѯp+CڍgWedPGTܾUSRs?EkfQ]mVV_} `x2W21\oِas%1$ *Ot4h¶1c:8Z̬Q^Z˲st:= +O7jGRDNѸ_.6Û'VK=y-B 4c#=l{>J  yS@KO<1opUS6U7|h܊l3._f=65@J)0&wcOĉkwO>)b޽hnt5gp? tʒt*!R*iv$=PJB}HZt*15(ol1b&,:bSh!hcXDtvF͠sN.wF#8ijqoWŅ.l,z=(BOЇʘ6.̬}0ͣFEQ}Ӂeyףּ1;vHN~cQWN2Q#!V_;\M &5Ѐ͢I e}^#/K6i `# 'm{XE ֋]%YZŚP&ha qP{vH*%6!2)<ۺh} uF-6JI(7t-o7?9M뭛n}rFYUzQu2W59c~A) 9. C by̱)7x/4i00ԸbW@n6Kw$LxM`l ,,䴭}Q7J^wWyl^!SA*GAX/ m(Mt`\EyXiLαa%TRrc@r!ʁk?i iBTfBL90#HP(˜#s4hNA:԰H9Ȑ;-h0eH2թi b&ZJɰTua{@MY{ː3k/d*<(ϫ_WܑF6G^ybP^͆5FҴ7z&rR1&gY~"j* sL_'50齚PNsi׆=<;*yla^04jgRjt4bȋ0UFslbxd_~[ju*R9&GVeۺ6ՏmۊS6!4~R-wOҽy^mOfŅl 1,rE~ˠt>:v#@¬6eVSw@ZrjgKnolW3jw3[d,VGQ|4\,zn֟^׹A+[[UVꮾ|u>EK [fFvAlj$Xh_JT'2?)^śœ7}7O~|F*xy:RbpL¯"{eM3?uŀߢ]-v_+֡mmH׽_f I廵G}ٲ;Ap{yb~ZMUTt*[QrB1wz;c\#]FwMҵ߸d@x.RlJQQR\n!.;QbIZAi}>waVzA=OAIQ^:rb"%h0cM%b:^%t_~ν|.$)Ʀ(R"7[$2htl Dv֞spU+?OO{akppJwkeEsnϰѩfnO5{e뾧h5I"-A:ڔT]ði$Z2g޺b} JkI,B1SD٤DQHEGy.8Yg'c \tNa&M=MsIY}G#ys.|$ݧӱ*񹣄9ep %Ɣ&ZD-MS$r69Cur`#e s;6V6S$h]DF2BDSٚ֓5)t >9I);"A ^* Dγ]1튜`ա{I l_vb.|1ck*WT❒ 1 9X!Y4KV+ 5e/);EۛSRRłPI c9댜[ٞw1[u }Ƀۺɾ 3uńaH(b,FgOB[BRT}-PԷӏʳngȨ0Gm5~6ňeW*˖rFe(:Zi:8I? Gu^$$E!@d)]X"tfU kMbɊL8:Q{mgldc!ݮ;V71Ệ'z/WB·ړ(V, 9!t#9a&P2Y< E0YhؙA<kb_3t\ޜұ &հ&p{E\i4"_-y'˄Cc%+'["l2ryX{.Rs>m䶺e4&/ ,Z$-Рi O.ڊeɑd:|ْWokSNug<<'υO™<)"RZXZGSwpN yNEݻ< F7+atL`RXܢf"锨@47JъǍ=0nlm{dy^M֏IK|+O4 yLgu[DWY eʒnbpgGOĎ8I$N bHIf,oghl#bSыZnH7W=էOv[lJwh}8= r_GV|Cue"zyq2Hۼ`'8pQ|Rm$4CH} =_- +V:]EK-Fխza lS)jϐ4uU ˮ6$꺫N[>S;kOɢY7,[ m'J21[q䙂-Jը^9T527erv~]7\͗8c^|梽/Z{\÷7M kC7fMt3l_$,-@֎U/tt|M6ͳ0εin)O0_5n?қ9nOKXZnz|[x{5/&g7{ .my[Xo/Rq;խ6ȅds4?/k0jNe=DW߮\3!/, O0 )Tϭ?Ea׭o'K,kN-7jjJ`}@ Aā5Q(,fPܙ䓴JCNL1, QR(OyU688L{cbT-2L!Z"ՁHO3qqCታ?\﮺.GUgv9]iexc2Y)fC GS^0>yFλDzNldB*E)sTYK 9,lœ4a8I @ɛPmpD>[R}َn< `lq_6Q3؍q;$=AIʙg3g 9ﭱ) !H`ҋ"{ԼxT֣T22CQE{b >/!#b.$EuZQ? = oG7=O@ut#pxDqHO `"Js3"P' ΨIRQq52JXޣUGrnˆX̜4vyL!uk%EZi=.njAJf4Hk! 6Gz xd:S .hQ=.!C ?e@ ~|*Ge@qn ;W39/qb9ڙgIڻ=R{/wN~? ؿQxr!C˅Yfז6oNzeH'BD;FRx0(XHO)mr128%j$y[}wP2wV2m/aHhuYnc\Ñ(Y֬斸:{$Oc㦗1ٖE=ֆu7Sa ZqS$jo8fMEAPQ8p'u&ƄH卲LT[ SBk8(KP[9Y]zkoiSl;$ېnq#'ڣn S'>G,G؝ &sZk-[KJ@hDIPk+(qbW@= QH18'-b>$('g&&h81` 3@>F/M!&NKحwt&語{-h|5j 8IɻY&2S8Cw"co/dn5AjET]]o{}4ФYީ䦧4tvvuV@πtmkL3G:@AzfO)xY<crx86-Ayi<9O&Yǩv$c -P@. B IU>>rZ7!Qey?<Á@Yy٭/MEʢe}{#jdj,SʤupNfN G"vJDmZ =0<0$U"cHT :0m21ZI aq\ᣈ+H+26rWz5E&Owo|[][Xb.-ʛu_8Sx Ev&jbW&tB)dbycd=dvzY ==4"6z)a@3ƄKJ#+J F-wFgrw#Otan'='FgQV'W"=#K41SXAPR&}/o'N.S5}vWQ8d2ͨ/"Em7pFK%@(֢ :%LK55OvJ!W G=?>"ovݬ^5[SNҰF{S\sЬ-RxeW4dT ҴNĻɶ.TrR} -OZ}:nSo-OyW׏_{ދz0ZHn`:jΧ@]Y/ -,,wx/W 8K֗ՃeWpg|hW/ (f^.d^Y W:z[p fpa >_60FK3vϮ[],]yW;@!ܰd7:}xi/E?~xZtxhUbAY~uf 7{'6i 8{}jw֏ /ϖb_So3]u:\:Z;] (剮>A"V=ؙ ]u;]u+߃{:\}UG)>AZe֮/tFc'=i:=2c2<-t]0}crc='DWxc9x|IOI8W~<!ǰwjfHvĢR{}܁!:$\H\T}iW;-*wC6I@HI i)}KC_p]1$gNQ]^]yg=4f"4֍^~Ϛ\ߑ%֊s?eQ"g%]|lJgն9oCf2iwJUfwUבˍ-\/@hǽKo($ z`ۆO/.idӆx1Kk_|}Mn7õ6r[pȹ;Pbs*jjsJ%7Ld0v/D29( ծiB:woCW?ԄÆݭh/َ\}\m|٠)W..Yx6pBl|C~+ϐwͷuՈ{hf^Q~?]צߌsK#kVMj\ih&s.Ew*5+]thWS{sc &k+ ZMW[.ƏWDiRU}p)W*@z([F Q |Woz̿o]'Z䉥6)K,modCk\qsͿ"kG7.(_<f4K~,IasJɪ2v4!n>rgɞ#&ODgmcv'lww;|0g&I囅A]~(}x&xp\ Z z=罿*޺kd}[y|8ϩgnz9ͭM.1P䬼k26X=ɼv7{)zUn!zXD6܉X)$;DS'ڔme!ݽкӴ$d5퇕CaȻ[].o=GvtL)ˡl/Coܶz;0J~d;ˡ}erz+]O^h䞥3.|4_t'/xoE(co|-WǧR2o^9M<ά##odo3__C[ף7G>x둫cp^oQo۹o.v8bMܳhf˧]ױqMW;IxMG#PQA]t[Fys`wH>!-k_z\?84g2+wīb8ew#^~n rָЪt1I$JU'u:ZB~AG[zĿHdn}W.gF?-X?x~tR+iQgD7_jhX%ΚzJ>9suetq7u뇒XLtVIB*e28)5Z45V|bt vbNd-ISZcPTHAKdSɘ DU 7m[ѱ(6Kh`E,6r5rIznNiiv6AI$jU&DjY{,u\xb46cF6Cc6 66eZIPt)SV5H)F幰–Hv/NedȺ0J20h=h5ܿPE6P&RTa*!o>+&Ul,nYG>ɻ0Fzzxd!aKhĉ#i˳CVCdg1Җf$dyKJ·Hr0L;oBнfUS#oJ̱6v0L f[5+A5ArMtF¶,5D75r'X-P"jmߠGTUC+ΰ)*2X -DmTЗ 90k]'O!B`QOk;i_6gHAgĐjWQ ^,Zrc`ĉ=HXF*H`jA 1 o\*KIo cL"_&31SN 8(<4&%NWh WZa}K :83$iD)RA!WOPb$Xأ0xBL@pOKL. k Ӳز6rqֿtXv=ZsIV֑ Ĥ bKH՜{'еMq$w_!)h~^@HplId3!T"9Yr9m.`3۽ToMO6A8̎|-! |Mo{ I ? kHI`}oC=4X3b@\:.dk)It:\c|%SIǞ>5FH8 eԒޑn?piað) }uƏӥ`ԗ[wtv\D&5=F$krX'o !Y4[B;cLd5_0*ըqv/|sqw}qq=aN3]6 }+XcK~3$۾kxIRI6cn }'673YŶιp[T $g'' dꆘ$@D&r`m$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MihI M@|I.Im@ԧ{-" YcB5ı?NѧvjoVG&zռ=B ȄU^]^\oJOOg5xd OLrQa4yP 8M@iJP&4 (M@iJP&4 (M@iJP&4 (M@iJP&4 (M@iJP&4 (M@iJP&QR rPk1 (SfM@-q/$O%I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@Zp'$,' DA1I T}(}$B]5ٸ=}3Ω#Џ7 }P[gG-rDc~7SJھ9iwнk ͯ=` |~n崮~;X!GWN?|nYwlyۿүā}_N7<4T1c:FtXfm/ܲ}!wl^77ƃ雜o:DG#Kh@# iKM[jRӖԴ-5miKM[jRӖԴ-5miKM[jRӖԴ-5miKM[jRӖԴ-5miKM[jRӖԴ-5miKM[jRӖԴrҖ/72Jy)~.]P/踻a诶J4NFR8M< MƛP}8(pá4DJ銀sbteqEWDSbP]-PW$6%` A s PO7+H):I"`^7{+^ SQ:ʧ9 ҕGqI~sbʠZZA8_7bڙ֚u5Rfu銀,EWD]WN/QW +(FWbډ֛ꫫR%&V8X9F\1BW[uV2QrtŸHڙrRXu%\WΜϷqq:mWn^gWG&1p=ӫPkMXX8gC;W/./O<ƹ}Mh,YonhOv#oΨ64x?\_vWMߒϛ|ノ g0zpw(w5`X{|͛]|Sp`}5y4p׽uk C7-sYcza=9'hsM#V M;uV8!zNbl'w:צyqxuX6FD&n5!:ۥ~}?.B>:/G[۽fgfs ~.x|c׹{xL&~~wrqm댍 7-t}7 `ǟO'zd;|v![!nw@!-s25A#su! ]0q]1^9zҫ+j82`+Ɲyx!^W#eV]-PWMF8]3QKO +tZ|F ܧ ո7c+XGW!ө* (FWA6uŔu Wvƕ3wŴz]1e C4$]1. 2-V?ΔY%*`uWl]1.x)bZ_T;Sju7WWes 2XF;T{lʪlcݑ  ,yͱ6L}s~qu}gCv&7ח6_?v\Ơ"Ul(JSzlor ]|qdKo k]wPM>j=uS7i`6&q&vwnj ְ8S 3RJaQK%,='IW ]1nB)bZ~ΔYuD]=ΪO*:&b:qۥ=XQ?;2h?ך*Sb^)hٻ^b(HW ]1n)bZ)S]-PW`rN8'1"8&e֮+Z! ]10F1bhuŔUW U`ꊁq=J箘Yt '„qu}AbjSNndI8˙j'`L3uŔN箖+h@G9sW;^ejSf]D]e4uW <-esoRF 2%꺫GWiϦO'9[Mˀw)řhL뮊(lU*UR]=`tkXa~eIC>*8n-ʎ+Cd=5|0|l HOe\ڲrEycR}S 3nHRJaX{)̔)j)Rع1 gSq+uՏܙtD]C ]s(FW (EWLl)Qu@]y1tq}+ 2&u,2.#1b\/2._]Yu@]E$銁QT;F#FWL;WuEXbtŸ f0ȴWW#VWKB$ {kqiծRuD]e@IW8YQBMWW#鵠6hIG&B^bCd>A(8no.Ge8w 7ϬZkL.LXiTM?驨'{N#4tR~RE{+f?~5%r?]܆i{V]Cυwϲog {|hwo_suKW#7M=f"}됚5Xl;߿l}\zً_}z߁dn"o~ͰW|wя]뼺6\mG}{Nܞ?= ?JXPQﻷCsvh?[7_tޅ.m^uؘkk7 Osv rD<}O*ٽ޴^TWC_^t4mB\N]ĔT;t}l9KYGtN|6_6&۵Lvf.[{=򮻭>t5Fߘ^gO{uƗޜәo[ҏg7=V뷻i/"'͠]C甦?|'/fL?G~}ȟYޞo|yvO/IqF?ݾ?v-{gzɑ_ 6eއF04ؙ} 2,kJܒ:7:,tEٲ'Q(+ 2" j{''^j)'id9&C^U>B$.<']h$*8Y*$AԨ4cܦIF7dNP!rX )gS6 cJxxk=@ds!ZGdrV嬕ϯ@ZFŨt)djG/y&bdH1(!欣Q` u9=Nk>@O?3 K<#%)r#ʊ&<4FI< wjUZ0t~R'-wMlýxˍ{˩!YOӿ̼J:$Ӗ#cY%S2%3ILJRv^kE,aSU7Gvabޘ? vO0f31jZ:YV"7䛃tಿ ~|v d!j0<;a3'wASg v]DfDh#sЇUc ,|~.|X}fce"&4yp9E`^%EΌMuQO<`ͧsj( 8jJDz Vn4mfdoHTc}QF_ۯyrfg˃d\` Q! 0 fi!x7OS9FJ駷W~)}g:^]\~$?:No&$U ?+0鷜MJGd;G}Y1=D#'|i. 4viͻٔX g}\#?RH6|odclL9$=g{fqT9 cgvŁlν98ECmMh3ta3עFUvؚ9;\~XnN.YjVƛq ܌G? (dAL(ȇJgYrK%L$rB5JLCvn`٠k]0z4`b0b}G{0,s 7Zh5nX 3~Sw1Je2ȔBK'ђa9bJICpo$/ϜN7.?[=]uh6΅y&iɢte vQCiH4cܔmhn"DrhRLvO=ǡӝI&I4$Áy+ȁ~OsI{2dQ+C4s@Ѩ`U\VMD@Cqzyh z sI&Pzۄ"D+'"T#pPHgW;߯{f@Rp@Y+nJ'Q}jl?NŃzl =:򒭞WzP;diԋ_į8<@Vj% ۽0 ~ gbhחƶCZσ8һS-i>z5^;8tkGzXHdSY:K= W%SR쑴ԏ)ȴFp̳dm)VcNlv`2 3Ҹaj "D3E= ˘ IJE!m2'r ^_Da*ઑs NZAoj.eƘ!03ňQ^$)[n[r[r]vKɷb쒱0me2ϣ>RsV,U&cXRs|_6\tQvmr8$ h˟oSB_ (a:k_pU &́|mwD_@ro!xAFAEKğQ` (^2U!y}eKYׂN^;9L3vԖMe4+Ɨ|kttf+rpuKx~4Yͽ?%l)i2Wʒ[bKu`+¿uJ%r'ine ŭu#q/m_C n_Z7ֻNtlPl; lzx|~=hGw܏'|&{>{tOY'/2o|7wXo}ߍݻu;WgaH3E83~1Wu> 2hT(cMrv úH\;C "'Gm,Q&'̢W* 3GeiA&4\G`9نg&6gKO#}q3͉r`g-Zp;YOlo>Luwȣ`^1!di (s J z΍ *8 >G1;.c @B.Ug9.{VpjyW973Wr_yYzyrAH3& "ƤyCYN"(UC85".Sա"q(yv \ TѲ 2A,c UŭOVd4T3¦Bt.í~rQ!=w&*81Ӑ`A0u` <0a uA*f`DnHI]iD$$VH-SDA}9 q`)f4ʼnek̏hUOH-/KM?RZlK$2K@D+@ezJFH:)oTՇY\EC&dy ђ!#P80IGхS&ir>F~}:; b<X?E壖4bwq|Iʂ;,?^D&^1Ŭ2VsC !B$ G Vs3fu'j%r虘"8 yBY,rOeXҧ'=1%Ջ^^.OƄ0ԌieGg|xF:d}VcbrFuzz\a5x>x>< qE, tg~ɞzfZ`@F Fc:(F԰L?#yQ."ʧۧ.n|IP9NYQTsEVB{a9ef#&xruӆ\HH0u,e}VRHi 9"FK5,j/"g?`|^NC9fBys@&4W/Sbf#Ag>R:v:Q.6^mƟQ3ycq (Eh gou@Pw!9q擲,ɞ"s1 TVx1U cI+p**D`frOJRD,rBdZa)AP5r ҉㲼Ύm4kKn [TYʿh|@f"ORAEI.yF ZCK s'Uw^ɹOѴٶdx6SSOլص 6οy&F ҿLo3hYKAhBlt7oJo@W&4}x=:NE#? dc<\'q yZjgG\)S2$QVҋ8Hơ˙{曹?E3B1CpCE6IKl͕l͢;ݻ?,A0>m>ݛu~9cxC-fKlv<,oG}ofnǡhVgM_/R mqVM|Ajq$H/Eο{^+],Kޚ/DΙ _oYbZu0g^ќ]=G}yv ) ǣg؝K m@[t$0$RIy.f-=%YW~ekɳ71lrnw%˛21ߚM_:g  3<uLυs)FR9Ž NE$!t$;)gP JvJlS9WB^@bYcf&{ K2Ij$^C; Y3k)nBH8gtVwP'&mH3ӬdXœn0@tr"8FGZRdYx%]BO3#+`%Nr6CN0Q-1H.vtEɤDQ#"r]U4g%&yp(ɤHXO@ϵpΥO *JJN΁,dsVKN$O. ?YzE:qȒs=Z,hg1h !D" &V5E:Q Ze 鹾3pLY/C <-8d0tQE][v3!^Bgn}hNŸQN!y,% $ADht/%ڢl4b|" ӥ" t" o_>x\CB #Urڢaٛ Up*mQ\6p-p6h[ GEq;})ZCEH0\r{m,Y|36_cUׅ|pbЍBy5kVDA=G[9~UCxHY"k* ߚA? a ч[SR@Ψ2yJP'(VDih!f8u$)+/d`~ W7mGq~05pȤKضVVx:,n Sۼ<  f]95gˬVolddo~[ 6BLDe{3]w~\=ŋ3^ݢYE3RhKJ*Dr28W!Zk#^>߼@8j#x:q\_?~|)ظ}Iq3·/, ̐ȻTEPim|_JtiM_m8<OCG7_mq?/IS(r1gPks5Lstͽc>/'m7GfLRSaV7o t-;6q;mzkVoSWН_3Pk7V]ztjY13]x7zͮޛL^x9ֽ\JN^`!hεbd:EEk_+6@Q{a:q,j|~QC: xlH"8 @#Zb% JCIL%j{s7B/N} 4i:w!.{Y ^1rOI}qG* 3䡲b9ZVJP& Z(kQy)%jQ**"3Ϡ>Πfz2w{dމ3*1E[SV+j SB$Kà LFRR?G?ӻۥ]VVO:od|#%Q**%股rܲQq[-#ZaƳd HMz,F]MM>fN]n۩-X!{5fՂ0q;1u`ܶZr[ZsJ!L":[_h}/wX;ҽu 7F8'Ѷ@*dX@rW܁pAJh*ɸR%훙/% ΐYk>4, {/Jp[5?=|$sYI\\l>=fpS%fL RzR0J D+Fx"8pswoAv™v/ph&*"k\ Ō`)ѐA1jP%kR$\ڦٌtS>RcK Xg\Fg}N1 d$zSwE65*D>&-[y1y9u Up:P z2@-cU+ux̅STUkefQ4bX/D2JmJW J1#Voưȑk#}-͍lة n7)"d ޳ѐm",d# =uyR)@>^Jwg7t)]vo^N\t4H8LGI !Q<v,EìNQgNp=:;{\VLwXwk|{F[ke6XweimBݽHF6R΍ֻ=!Z˯YY*r꼢e"" [I,X 'zV#+V2pH! ha%bG#5QYl 4>IXN<<nԚoȅƉwϹIQyD[h+ZE 3r3%XE,an+\_?D&`}UnKlXK=*e[sOAAES_0Q`'P&α] ģ!-o9T8^ ˠ+jroiLkS@:Ƞ6T82X)@q9 X% >WxkynuTxe(/JQT-xIZ)T,4ĥ *8|ARes~?jچӄ@bM>jtlӎD|+& AcFš.Zr} ^R2 ]m8a)qR5&%TC(~ ~w!@lha] Lb1 pbgB)$)J'W͍23&.G0Fީ|. /^)ݒ'5[GjQ(4u5h d aU =;8< `6G!xvC[6ڿFpwV SA-Ģ֓+Kr9N i*Wa4SE,\1-W fpfVK-]՗s DNN>[!=e\0|){UL5gBi]ͳ{ Vnw^V7]O /Vb钘3`zn n/,G'/Nh9C=lI# ~MðY^##F+5+gM> ۉ]98̯{8#6iMϊ%/F)hy$,q8b+|Hh1>c)NvS܎o~x__?ͷ?~Cyo^pP1l"A;;`d~C[8^;47q?ǘo1.]>qװĕA;KZ1/3wk[BkAw_.b$tQQURjIU81!vԿjc\k#]Fަ}v0D4Rb `Tn'ATT #kd)m罝V>:(i)6e+` 4^:BOC1.H1=M~#G/Nʖο5hk\ ݁1q?Ywnё&)%Z*tvkJ4k9}O#OqDns _d*x,^~AB$@Szrn ]Ki#tQ>{Phmqp'U`b8뺭.14X"'s DB&QtHd`4 A:# һ&5adVQ3h"4n$r$nX (^#lHCWg`Y'Ԇ;*\JLC"-ԫ䐡~b^sfVRc=wy&xwn+4>b`gd?݅Pv{w=c"uvQpav^K-'JD9Nh\$x_&j)` s,NtqDvc컻fm/ex?QppTq%Wx4n;z5/'m;.RvF ձsM XbP&-~ ڨ<в%FAm+\Ru*6B)J{(,cFe _$Wr{UoZzm~ߏ6L\.'Ho2fdkmzxmSC']X|l0mV@YP@5Z *!#A*!JCS?M7θJO 7ayWya4A<<(p7(䗤Nw\BKυ"?!q ~B*+SWWJzq+!zR*,W\ɞ**Sd#@Pb٭>^' ЊO{6kn~Y;sL?^Won/m =nZ{|{q/>ݏo7%w dMe+WnV\/ EwStug\vM.|w-z@c=;nVijsNJLx?_O;_ćy[(h*' y [dAhBގ-Bjm-ѝ_. kV Qx{v)|Y W+R0n4p)ء.EDW 8Mz)tIc (tute*,?= .uQz/tu tIo^⢻7fpkvW_\pDb*P'ɽ%~|@ؘ wԫ=y^{/7Wo=9qvkHu7'nphM1>O.`-qYG(~mF#dGG1GGy"C8{?[&qqQ]<M]mf{ |!6s$oձ{>5[kmf%t]ջc־l)? ,߼g5w/~/[.eoFUoz%7URvִY!נr6t>{gU?=5%ұ[R rV妬.R͚a¦i|NǩX4i~\h@>TC@(>Y$Z@oCv[E3C$ZJQ+ω6{Gb:0JHkfh*gs#QІ oT#F('@h%葤*߼ȵdCU,Z:d %SɎfT}P)}z8oNB<A޴\s9 Sf0j$7.(]cpOp_c;Sknws5b'-ؓ9fqu$~`یU) 3!9α_g~^YbHuEQCBNFG6Tق1K(#mtAڳZaB BQz `=vF|YW>BZ%;6:sP[r=JNwh A^Ak6_ h;oPR6J2Tbw%.A2hE ǖژ[Ck+.t-a6tGV1- YZec`T" R''XP. )ٰQF *TD@5w˽AAQU6MS`N9x X5lGD&0VN ddR!}EyI6>*) vF? VԠ u|/r3ki@e 6o]0s@ JAU.1a RLAN N0t掊CC\jSS5qC%:sQ%@ i `ɦ= p\AQ{Se*8(@Hq`Q ڳ;!JPA/uk846Օ/H!8I]ZD5=( E+CGAyǒ_5N_*}AI8(l,@|8=(aL! #A KCkn R5٠db̠JUYa:iB0`/s#vl44o3?8;5͈ gv3x:p4@.AϛБJhIWo#f BX9Y4<#y0 _4+Xq=ë?mŬuԆHT*Ǡ}TPGQpu $$7(X1z,1I]K,brjFF^ZF!Q0MVf՜%C8˾L <^c~˹+ , BttsYhl) l\c Ft@(jqHuutk昴58)knĘv3@9p gid|&% xKd|ki樇*tyF0C{_s":Y'%\aۊLv0X$SSAv => `yÎWVh7"$uSjCn ẟ,;:YoP ê Q GʅPeQERPGq1* x<RuN` tB)1QFǤz54 hc3(Vjҡƚ5AUJFm:?3Ag2vT VPXG%=ݳjyAц-\sA-4p^ kJ (Bu(AdJ 7bldNҳ֞fJOQ!,yJi]0d@b~pQi4 65fs˥ZpU]CC,Z*]hBM:SjŋPpԅΝqf@Bׄru3B?]wԢ`|J/L0H5vKx!/n-n; IBuy+u<_+Xػߣ8}2!&3$۝;|i6}a$̣8;)xv "%9/gpb|֌6;2iq[qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8]'RKr08-L ` tN rH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $Nv^\ZΙ@@;e'):@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qt@В@l\'XGJ tN`!s $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@:'_ֻd]Z̯^^RSnfy}_\[P?ݵv]wWo/o` D{&8rKc\B"ml3r-1.}ƥOu ŹZ ]1Z6( ? .S\ ]1\BWc+ qAtՋ)c+FiQXkvSRoo/6 Up_]Vm+^x ^M2%&(^wwD?,aĘ__]v?>juthTri{SݰWo/0?g߼9떑.ͫ?h?ڂxo|#in/+&^ׯnٲ[(yQ/{Һj2%Q/Q~48Riڔ1z[Zhw畽M 0fo1^.{׉]]1~r `BW6c tute/Y]1\KK+FKQ:%tuteyGH u.b/INWҊ:EUt +F-UK+FKQ:-tutRp: Wbz19zڠLBW'HW>%t WX7]1\BW (tutrjI]buh;]1J/ NʷJWIQWi1mpNW27CWC_Va3_Ά_E0h|y& 9`%hsQ glf/wxmE[;lF8$?r32i09C)#rų\=)Sxw09rpIv (7X+jwTsVXgVk E&)( =(-K˛N^(ڶ[Q\chH((jkEP5^/=*ύKZ,)-u[콺6SJe\ )eU@tʦ-slZ7S(YBVU:VpJ 1[a:[Zasʄ䊀ND*r(9d\IL$$Wlґ+µ:\!eApP2%wE:"\+S+uѻ+/r5R\5wY2rTh]Rf\in [.+<h]rE2OQ w䊀HFw¡Q,Wc+b k\1Z\YF(WNHRz2H"wE2wF?wEYR+gՋ)/C5 ~2FkڏFi#sW"@DgWԲ1S;bT`ܹa*`mo.K!%p b=wЈMB,wDh(ËU]ErӍykmUPGKm-o9sMYqzT)xmj˘W47BK5BDn)4 ei_-vzzr#ݍm{>9hq}A@Tr1w_l}:vYNk]Ymgxt-m;0]Bvp1t,*++Ι- -KWn }_)f]دO?]%s rPE :\JEpu:\# qemGTՋstgPk7b9+7(]c|trtk([՟.q+#cONuS}`v{>r݃к?IK-oW7{naɢד9:>zG7 u7/)冿@ 6Ng$`kEGlQZTBNߺ=:._/ml7}ܦ}=y @4=dr[CZLn )xܚ9\I)YBO "K4, J%$Wd2rNTh;Fixʕ&!"`㒑+µL#b.v"J\P4p)=i$`Eъ2]Q HjatW'#WDkwWDi\Q,]IHئ3wEΦ"WHwWD)ʁu.+GIl2rEC'r # x !MU( eBt\,WϮzńvF+34܂عc*ܲ6|Rzϝ&] r.5H4==XP;Ur;M1͗se#ÀHv.ۋn(=2(=7[tBr|QaC U\YF(W;0`i+\~R(MT\_&$W  WThM(-r5BRBJFWA*rE.Wr5BBi 6%#W+qWDڛO^.WDr08F2B tT i]u]QZ4wE'#Wn2 Ŀ(E~28FrOiU:rE: Dk+y՗#WfϪ7S&/dC]h@O(]dredzvՃRɫT. b̫d`Bq &)U d+9JZ8x.g3)ZN w@ݿ\B Àm:VõXa5,z+LeVMD0\2rEڡׇQrjr%+:"\S+u,v(mc,%wEN0&㮈V(cMJ%$Wlm2rEΤ"WH+8]:JjrMi =a㐌\J\%,W#+ڰ 0¥ "ɅCh-DxʕO^*W4*D2rE"WD ѻ+Ty}rOk!LFW%#WDT;Qˑ+g[Z//QsvWp9QI܉Jd?GJکI)ӄ4_EoCwt Кr5|0`!喆mo0 0\50^mQk}~ޛTV_QDQ\]6$ěoǃ@ͣ_e1nvYۓrrr.%}}rutA8%fkh SXx\]_Kn|wܟc~xp7owX4]5n-hͮC = ~O}s5_?Z G?,.}~q5]l{O:Bzg鏟W~];z5ڇdy1ns~{~wwyV,f #Ts%N[ע>Fex顬PLwaVUryv=?ʏd4h@E㯏eq SV` }uZ MRtoXTeZ\ \jU V;km݊2aaazË^Kj_xx%EQ^x=kڪM¼a{ wbʵ8|{QvvܜE􅻃B5$_OS' VC󁧜@g{Q7l/}3XɁWŇIWmݝ:*wr{Yh\!|5mvַ|yq!Uo`\뭉~u jXҡyYPUl& bYˮ˪_b3RKOl)4 !,@%a_᱙=Uc8g7UgymjGt,Mi+ݺD-14^i.p`LK/et2׵Sd5 <:_[5lecfO8-Js0v,[l:|LSq[uy{1oU-?TAE7tg>BYɍ(0qiTk6 $xW|kJQcD#ic%={XCw~vQvmL =6 z%v-G',mջ;vt t9wvvfBMEY4/Zu-ªWV2eeɈ,zuj)iIKGM-UVˬ TZzl*hPJ$jV(c0\H0x@VP u-k&`8{FY[5c"9ChaGuuќ-ݓOO_=}.[?Q-X^ᠳ=_oN>GY'XF)^J X/ln00YnDá/kAQϯVh5>ŷFiM-%96JhU;k87#|k`Y^\h=^~ژzsgYiW*mAk<(-)֔u6L*]*gm)_67A?m ˭}:, n=m8U48H1ִSFUSak Aj rG:pGz΄Zr^6NUhk]UjK(+Jz^SUb5Z7V hKY{|BM( wQs?YT[sr?ۙT3yٻ涑$WzPb+׎i;홇YNm d>YxH"e$RBG /deN=_$_IR$,eW(0 $ٟc}M0ƅ8JQ?C ;e̥ܧ[y*Ap<0_PA >ug e)sIg:gRP(\Ug4脞{ӿ@0!Z. tH\c% &N͛B1.7be6;?/|19ap L $,O+hS_i 7긵ZA/&901]q*$ OT9L~ɮnUSx ~aUwޕ7^OEl 1 xF̅-b{v^-]& ȋsS&cЪY{X Yݍ+YdO >`,&M/zWӉ^tw : uitz]5V/>uҧTFR>~,Eqg*K EwTO|^-ۙ΁ϟNߞy6?Oӓ>95h? YE[Pd nu?'Z릺]S~5żQF79.9ҡn-/هׅ>\~[+z-sY%݅>GlYW꧛T1Qm.r6foe){}c#]&',$D"E9 ]$"e#mC Vh)k;cp"h,)+F,3ޤ9%NeBpY~iN%4ם1YuMy$3 W\o(G[C&*μgȃD*PcqQ\DhK&C-5" nXf3װb)̘g !&%& qR49e833ՊGs13r\t4!IqK}/,:%Lv/&y$<ڛ!2Q m%``~/LswCSK79▱*L10 f1OGB[G‰uXT<xW26:}HIUHzå O=c|njg>LLj9^1Tk%imRI+9&5ܰZ=U6e}}q~~ӶRoM2qSb҃۳YuzMM^ t*j19!!ps(4nm)OC{O%T/RxCJχָ|0zXk~^J&R/ |Xxd!Rlb)#" b FрG!eLDV󩾟yvqYT= օy"&QWH4MiÓ# ,["67eOT!Sハp}D.\k>ڎ`)}F/ۑ#lEDa[Rnw\ DKf .T>șvWGy~6\Bi…A $a![&uv궕‚&y6-f<ܛ/A P|\GRq1G_˨ѴBMfUP OŨ# ;>TMY`/7?ຢɀd8# oìu/#.ϧ['mNCVb}֙*R׬e/K`1co7,쑊9No'TJSԀpBG忋$FL:<E[D /\)-<.S@9%L[N5Wbz9HFgDRp3M<fRHp\ XwL\NvڊFLƝN!9q6g."KL檪* \E$K8@W` e\E(GoUOk7v8%xv8LJr/pj*Q)[zpF~c9 JR/p;o]T^ \q' f \%r Jҝ k W/Z7d!$bWq:\%*%o•JIGpkB \%juJT6v*b*J❷^"\iI~h{e0&ԞU;JT6|JFl[˞LmlgnlbI;0Q&Y /8 l7Yuz{k\Xk)ZF1E44*\h*}-?zMT$oLrJ͙"ǜ M0&ZYە}]yt9w X,D6-ʠ\|ɮﰬͅiYe' {t²WJ~ď8=B ya\5=cc!vDFs.$V 6d)|?SG|bS~j4}h!)[SxӥRNO>[j=Mpj9^R'j[hL5Pm/9[fr;1"h "@0pe뇇n&&J*|EűKDp_Kk7/;=7n(W =ȨtD`}~n0euړȼ%~BUU?Āω4gL\ s/ᢁ]oHW}9\~ sIv_.lٲ%9vrI=(lp1ؙH YzuUuzZSGKz^QL{dL~&n7ܘx?kX+a5܍3qGӉ栺6>'Y ?SSo2xU>:ؓyG%g7^{5@otqs:2A.x*WBBj_(_mW陨 9 +$bJNyr^ NՂy7!j &T4e[CJ6=SX>=|:?[q1rfQp({ x@1ɭGulx*S!Hy| ʿҒ6,{dl7r2VQTnͥ+}(Pulz`tE5h&Rir:0Ii툋BT(.2[/.ەlxAcRF9z&A1)o>G;g H<QET$~Y%pĭ5&szfvqJ\-;\7Q" j9=_qnyo~ >ڐ4&islC Rbpht b7wۛh#;p e RIXrգO]8'd> xd `fHdFf+$t5 ,r>7KT}HRZhȾ 9vHo[eEnD,A GTCRfķ T2@|S 2s~ İP[p$H}#/ؽ=I{nܨ竬"m|>['0|Qi} FAEG){Miԛ~~ϿpJE?cݼ>N~qr&>GUaj曏*1:rQbCYP%$- M $ *h61BM~K^ZBR*8Ă'\#ER%Υ(xDHiT--r\Zd ~| ק#\9FC>^«h톀XksrFpZcL%+q"0=/.S1f}"jXNAk?808qsCq ŝ!V>jIU6^r[MhLvl+?O@PeI `3œI4 sV%A $X4pT'efğ'cfWdyiQ- {xG҈])mWcK|\]~0OTI@s@BtrcyRu;sy!2Q`>긿J`Ui/vq@_樍XyXx%/0\ jp}:%pi8g%O.qGoo|V/qga|d }VᨍD(f8>p܊%.>UP‘Yo @?*,s>;9Hձ2(%y0svbHBX;@]x),l~Qvܛ~nr|Y-Y]Ks{xT&SŤ`&r7k id̵̓pqz ;~37ΞKokϊ5R`~PD!cP+&1`yJƙ($WBN26*q`D!K lym͘:[ҜV-ޓvbLa)",5BC9B:3Ta S-8U8#'Zk xɋ~4E\M8ʡgJj+Ȟ;.?쉦ݶuN%8,߀>ئYD\Ojy)<@&Ѡq|*7hz%Йb3̱Bsp(?8d Lɢ^IX5ܟ6jQ59vسrý#@2_qr@$f(Z?ζ4r&sp 039q9j~^쬏29 _ܱuM^Jv~bJR'6s>c;z @M0J5Q/UCY;YS0]!@h)8H'Ǐb)wdv8A5%n"kRct◆i.Ϗ? S^ÌȽJtA(!㼦FrVNt0~}+sPa|<dz!4i^]dW]wطXi`!2e'c|x16w^4-/-(o\ `Q倄IL)RJ<5aVB:L4sdT&9Vܯ0 U}= AĐOt #,ecj<쨖/C `.p5{u;b'F#;vGmDʔ:dCqp!k_8nER Ox1s8 uEeG?T>D`|V` Z"/f^ }Hy"+17tJnyze"8 gr>9 (~^\zJ rL@6-$B^(> Z~kz sDPvcΞU 9JA0Ҷv M gtrAUʓ|:Ar$]DzX`ZJ"XuJ0/L?1vɥ@"A-Z~6ǏgO˾ qߠI+f<)SܻN_paYEʄL?$ T&!k.[1:ezmª zԉ-5u G$ Licihd qo[AF0cպ{! "B8"<Ϧ@Ju z?ܻ^Q)ђ6+0vWTK΍8ȆJYKQ1+=1d)KKZ;0*ByKJAAd2H@-gDqy'6|z* RVR*TB,;w$5Aj+_]OuyҲs'>\҂)YJ*pBկ)nSj&T.e~* $R$O3 ٱgW[$d.JUa:G?A)e,T„if"c$@8c(m1 mD-"f2!ˋ~V7x*d]üKOEϯ7Vg7z1?ŴaZY|?|s ^[瓃>O>:rwt~^]u^2R4PJHsV\diNRNPq8h$鷏ڧtoI{_X $ u'_ωb\'pdYH12I3@@ X`ӗ$!"wVd9-3oM{gwaZE @0"ŀEB 5G]L꼍wsܺ]̭kgnv;޸F{Qwd62*BŤ%H0 W\#c .]c˞VOav!~0m)3hAL͠!f"qp>^./ldiwY>*>'õ}:+.7Nm1cڝzpgf.ɱ4(d=B) BN)8S1s˦A?t*w!?3JYDsmѦHSb'?m$d}&wu5JLۗ } @XQa@J n!?+iߝĶnMG"k7e64rP&HGT@gZB)$d 3(hݡMm0r /58V_`,`Lmݓܱ+g9dwpAt;++0WfrL{1J\e'' i;1g>gq@{ܶEa0YۍpѶxo:K30+58cK`0&[U)DJǦ_#B^˂F196"u蔦w% zQd}lgV<"]Y1JF]Q˯*#8;kH6RLՠG"~.M+zmE*^ YG;h*(2n@ow$9,#ߤsyPN6PÁ6ՍNGfg^Ѻ"h2-+޼JB*d E I.{.`^0<]h;5RM/յ;SJS Yrju=IqOWDy?_dzA"@g B2"asۑVw}GnZei9#2i~tf-Y5/$i܌ J -9Qѽ7{*a"I6WiwΓoÎh}Î&~l`y &+&2\i!2bpjh7pWTw/QY>U{^|\)" +0d%1,YpDQkKD-4{yRfU3:c%69K'VP wǎ= ةֶ6 Vkk2)IiX"7 8D8dΖW<8;~:*=K+T@,ݑ/ $>Î@)BlW7a~ \ڱzSe(R(X3< %xnf(JQ8b`@,2|KKDIY,n͗[ز ofel Q^;vDx V8! %Џ(r V")0A ƾ<[rpwH>"ZQHFbproE{AL)fV \@DPt 6iP %(ɼ>qߔ)hy8g1OS턂'%ء|9A[ k d! mPߚ lC)ԡo^9^#"ŃB"<[5mve|\PH;|N!$b z'8&^~7Y Svco/*v~cT> *7&vKB1k\sǠխ+S.OY@(\0.6=)IG{2[@Lvxg ԊjDayJsӇ}.$^ݬnj4" ;…~$e.ӤKF1 8Tgas8͹x9)>eM]BBsGQ-g!튞i8 .UX\arNVXaa:E2jW<V%ZO*Z#`^ GSm}h5hĊib}Y_O,3Jϳp6HipH} fAmpBmG QJt"W}qU4 `ca[["ؙ`u+yWUX!^@;H8v^=aw˶1 5*BxBm7ǧĽ{H}rc 8}&,Ox]mgUy-X}W#<cDh(7$@jf=}D}7†6n?uM׃p"4: ,>|ɪNK,EDԭN[h~%ѨQ@3 }.p3UŇIgm}c_&ĥog $:wއGgӦK^IۚoNH<+k6H) qB uƽj! t7XM])קAλm_-,7cHD_ꛢ~d)|s^R_k;g4T4Fc\Ѵ6 yN`䌤޾# {[Lm,̫3{Sd˯ʼq:\2RG-c\.RLc\ek?QN2f:E;$25!aB.ݛ#dLh "י}#ۼq)_w1:|P8(KBI0zID<Й{},b.  eo"`̚=޲1YVe;d:[|4Jn8/(<;}eܧ"s E_E]DZ"/afמ9Lj{1ҍ0x. 9{pNlK(Y c8L oNEdg}Ǵ?:%lCdiŒĪt>R[;}S7W^itA(vAGos3&1?6#rFT!|1qaژ^|:ɥs;Ak_\Y.Xm2AOI8}G-L󳰾 sc>:Wz #&`U$g0}у@I-$Wg!(ܗU9/}G7HU# 9O'kz707NdOZi&=?OrvLO ]Q]~Z,Չ7A,?-H!zQJUe4MjxQq?MU5f˃h0ifMX88--7pCzbBVKkoP5 >B#RybgɏD~Oe%'D_jA,Kf۽TrjA`yXL1y,Jnirc δ^$b6t5D axϗܖpB`JA,gZ!M;uYE!ܽ @nmwlUp_{b F1m#ނIs^,-HfZΒkCk}7U(,HWctwJ *8|B8uTn66ּ&uTiDqV͋(4OdrNeܨTĂp$' lZ-l#=Xc\Idr1' J:.uI- QH5!$ "Cjկ_MFEV6d16he*i0l_(D*t +Je'4sc`=;5l'>>yz_D1]XvovNpFvpEy(vY)][M;1)nh!(-~\b UFJ2jzN=p6ܨW͸-aqៃ?3wY4s(o*4ILg`͖ DT >q_x[UMҥUOOK>r74d63l2GY~`| >J*P/m/h2>1%*_ٲͦLS!Q^KF$G]+)NC/*Ɩ)yPXr#ꂋ^k5?'ˆwEuW֞%q[̨МNFe_, J`K* "f/ʬڵj`Ñ"R3h -"~W`2vu:0C(aʗmdm=ЂC0'Ё לt=B}-JМ (!Z4kӀnyvǭՂidշ‚[ V + /wuvH9mGc NNMfB;NSZ(wʱp] qxЏ9Kͪɵq-e@)3D|jlKhuD2gLnfwJB ԔOXc)e o9L)\:3*2vc` o"7˨sWP z#=yъ.I^bv(ڑY9b6%\Я"9i&oa!Ph\,2 -cy.H!oe!;-c]m.X(Hg}ˀ=.ThQe1.q"0 Q9뺈pҤI2M*H77Ƃ64Cxz Uviy5G*.i#!2%!S}ɛؐjB܌Yg7r!W57lzܵ?^Ql}jFbS-Jnwm so?G8XYI75BwK%"k}糀Y,k}yd9&,4q 6x,H’sfTQuW(HutF{sƠv0J?~mc#[񸀁$mEh^˯l2Mל0ȋ!Kl9텼ڟXC~pT`Ya \!@9@1 )i]7CZx#Bǔi!'8|8xX;np[ 3k,iaP',ѳu>`@JNPUG2)Jxf$'Bu0u9+J 0n&5ܘmV;{sxAm,3R*,B@8befu{SJqqJ6Q4kyHkޱP* P1W Hh<2dٻ޶lW -#TXWnLb}g0@YOj<)rs%,g\!@scD|u;Qw%ܚT ޺ :6\:K$\?a87p.!b+e [#ZJ'P<>䫸s :C2&"loIX9d"0Z-'``혢}U`| g;~v F qgLڈaQhTja9our{~ Lc0?>LE>L Ln#.Df5rW`$~LK@W^1A" {eU0X4G[ o:0)J r'sa̻!"; yweVOKڛeߢDلX©p3>011*ӓc` (TZB `^AipkҀRw˶N&S :W{jibA ]ESAuB9#<Ћf]j)^z?==<&v86yXb&Jc v28'p(HxH+[tDz%$\#9eDC d\irzS2OaeZ46\޻$OV&)D*\Mth g[I#-m >VD^rD ߬o}ÆIr# (mƱ39|qp kacqWv}ٻpszzÇT`7 EFRy _YImddF%@v[xc<"0_"w"@]STV\0x^g3A)7J\k'+Q4D[bBbb! ֧4z[ssF0gJQ=jgq(C_ K}}T2U r6̾0svXsmD%qc n`Կÿ,vFoY?\u>BskO_kv5ooZ9F1vB@fg&/qp5RBe tPdݒ5ss\"i eJa5fH4Q01*֪x`4FdY2l އ/d&!r bG` ϣD5s,q Up0vI]6Dgvo* ٩ۺSS)_xtjVZT'L:xoOb598& ٩op՜S*譙~:\^-DҲm E&{O!? "GxBNݨ?&,Sm-.O qAEhCcQN& Qs* C+Ѱ2@FKi٤m ?A@t[+ ʒmRj[댵jr63:Ym"YE9&vY`ؐ+- !]Puڪ 溣1Aј6 ፞5~Y׍4![ `Va\ʱ 3wv dDiHZ\3^EaY+㎬vZc)3$*&!8n7N9lj%"vʥ@26&$rRFTJUyϟ)xT0+R ,`~89U^~Ֆi:hDG|q %p=O@;0N6leazi6 UAչ2nXWet'diۿM&ޗMlحg_) ~GƠ)'±Gb"xѩ:$ n8 `IJ𧨆0LΪ41"(_<Ȧ(QI…K1NYk Gy~ΐuU,\ | ^ݰ3HsT I(ڧv.` 3M9#qλ1U?"XCGNR#(Lq_F H[O"P P~EGQٴ!D4 6(G BPSXlanɴa(V,BV<\nj(֐u j {::o!0|@ߖh.jE304W$ .JmpiPs"*ZtXΛ[(a|j#.Df5X4VkH5R oe xبpRmqj\j8(S dZ I\!A{u8\6p!E6[t BP:rm"em摃Wa]'JӻM'þ헙Զx;T4Wc $ýC|D LKT@4oO؛>2DvT%ij:|2s~܂ d\?,oɇ /ݕ[ )n;MeJDلXRqh>g=g *gUvq.TZ-*bJ5O0 '**jU[WydNL!~+ΉM6ٝi(kP2D(REw_%mdz[7!ČAb)bس&g4)jUYxzÊӡ(ׇ4lI-7,%xIHy&@Xd2 G,HO$X O֬Nl߽&e^r*Z5zsba3b퐐_P/-f+>M2mVɁfp6w#ġGdj]*-_гr2svE`OIԶvfwM>.i+Ȅ)Ʉ`,ڂPE}Sk!_M%rW|g,MͦTrմ~.[V1T?G·ֿzgPvY$ *ZwPIY*ͩ(M)R"/۰Q~Zn儓$Rhrݨttl/Qs6UH:wef4pڦA׃5(EϊӨ0BJJ!ϓxJ{ִJpѤx#G$[>cL6B}w;W#?~rLȷt_vs/ VpbĉE}b8\hp*Y"m[$s"7TL04 {BL "Z>-5+῎n Ұü-ϕn㣳pYSz3 Ndo?]Q9-Y .TEz& ~.' U0%"Y2r,VD`OwyG{Q(ݴj^kA6\N Pqn~`&&…Fl:D呗}<Tr~C+qܧP(ےIq 6;,m"$+R.k`|cJL476` ]ڗ%bZrUR$Ɩm&Do:Tnw!`]`ٛU{ӎo$<9rmƷn7KN^I y0粗?O \o2&u_ QN%xT7JnQ䶮m԰~蜊 ;S/ZR{J ]OhNo2$hun/д>yp.!ꙴ˔] K50ޜdzK*L'Q[Gi_bF~UVg#c̒5K&)I#҃A=N8`OWEA9hK7!ipFTIaM HaKdNxh 8ͤWE鶀Z n |{~Jrx:feo|N`U8dIhj( 7bbK˚{W>ܿu2q-ѦW /G s XZb4-E 7/^}܌|y E^4a SIXl3Lz~ b_| 7%I0xϡN3E;_l_PV! XtpN'p#m @B3m3Mˌfzŵ=5݄6O!]/8qY-wn_tMԯUBh:a49;NqXe'6jiMr=<`x昒hҐ͉bW-Vb[۳QOҏV~8qB{t -k`ԾjNh?E{S@{sed*_O Uq$`3!k gVb]@k`jr R pT8297:o..+XІcßq&(9'̨PTϹmO?n2p[,I(N"Bq,MJ%O]^ .͌3F #dzf7@ 5;8%0P(&뚾5oh!# o 4xcrrI (/oL7="@Eaѳ>oۤ tE:W\^/$k&s`b2N;Dfl۰PC`{o&V@~)#i8NAA֯+P1Zr Scw_ULl)04IM—6& QLp6C~_68&~(k82{Jf?>X0{؟zԘQӆTNU +\rQEͮBwpƂ/uH3SadKlp>qyh'Lӟul.ܶn[:CqQ vQxw! F xVv]p0x_0 @'j)ADJ+B .3OD:JQjUbTStFscoI"9hOLgzHr_Yi1@?M sCS-kcn//C֑*$Ja v1A~ K\m"jc'[wJIky|[:oU5\ŠkVW9> K;FMdNF`WHV(_Ow@SEh6ДJ!EMt*}}=9pXx DA3+NQZ#c,f|A "nTHЁt a*սӢKb\ssUMy;z`%{ڱ`/Bͦr[wӠ)y/.߷yv/#kFʞ"WXLc,ڦZ'n]Q++#ɮh,$`RNaغtaMjRLLa pkB[UOhAۄ`"JCIUs C8(Pg[Jg:/c }Hg_s 5e>um d~ :A}798=$u͵~j*2Z9ȸ=Nwibh'W9|8l#;FBV<+^h~ӂCn}hK$URJߢyӜsoAj"xO5H$e <-;FC8?aɼ6Jovƶ1A0nqA:=tr 6B*N'HVrRC[w^!PIɫ .̉xZCx Swxqrw HlM~̦x;M;#o1+"[h"M)N)7ӑt>OG:NGCz)ڸ^b8!wM$rGG`j &;2o\ѣ! T 9Uf77;ޠanŠQ%QMohC+ Au !FLu_xثzC<|)ɇ=0'0}xH 6a'_ebAK!N6+]P>xYeKtV"b.JNAc=;\/hEA,r{OW[Ɍ.CFO}y^Xә~[t>I{/ -1 /?l Z앏%d!"טX)&rFBOUFV)$/PPv8)&,G ީϟ` dh|-Vջ+3%%\ A/q/U.&"!?ӲV:AwEޢoSހFnӚzʛ𒽩_#X68WjsqC4W_-8`݊po93W+ȃKogYG}Qw47@~_;RWSl,p;{}(fo9BBSMO#{~˫ 8HS@ր+x8zV  +:-hfÂ;,3)/iDP>-Wwjmn17nQ6| 6~99;rX0Fģ{J̤ %8;7ԎrhYMA17UV@\mZ½1uѾ9Ft.} qٷ|VguQׯj{ggq3FCLodd^W鿏ߟǿc `9\'f‚Y乸G)H/ -&vw;i7߶79`iS ZG ^#rN6B2hFž~[ yc,%,(>/b˾z뱸&o[=hj'7y0{G[lf6ZҷNmG Z,'ߒ#YOD@Wo(~FVwjmT>=zOh{XZθs\q7Z߇=mMq:gM4*g=%"=,½7ؐeZ*w`f3_kh(.33 V;s}/+ jF?&Be !%;6iGP1}^o0;g{<*~&ə"̺upȴGGC)/p]7+ vP6sl3gNVP{geh3g[c߆Wfcr6YЛ]<`qc۔̜mlw 0Br_ ?rSpSveO~BRأWw ?~=D;ʛ[)1kp>Aߐ&ܸi\\'gw7p55l=5R :CnZ=&03W * 6J4b-):&cwIE펙|9FYNdCͦ/A+iXNd":Q"jvA{hw4EߎR}$#T #%9!-q$( Z9bc.D3OT5I½Lkrp0seъBAךAQ9Ҿ1۝vP¢mçz>ކO qSMq:[wGcJ d, /(* sP^T9b9Q&l8a$d,@,3ĝYZЃhzNɂGLuvp0,G`IhRSx|>&-Q*IS₠e"ڍ&Ƒ o#h=:`r$"tɔ-6WC)Dk Sn *N&Hv/o/7=9g_N :`VZ"%ueNbX(* YV>'R !*GQGMl@ S4@bܚRrr%6#͎BQ  sopچƄP&4`4 J^Lf_M'MѢ1s#g?N#~9!6-98W7t|_qu=`}>@΋U|}˛I(/ԱBbxOZ\Fp, ? + (\La9qɖSĽ$euGvfPr~ST;4,49>tc>}FxtV݇p>PG+I}:r0%檪o5 N^e(PҭQ6N0fWL̝XZ'Ԝn9 l"c#eNݺvvAbNR\E%(/7!z?1>Ƴzz5Sa䢒eVBSqVy-Y9l!ʉ Eg4a;dxOZᨣ,)8`N{?Hb,ڦZy S،$Scr#hJڙvBYAX \ ىC%*6%T'LݔlQɷy T kkNzYLv0\A& bB-15ޅfJ& }A'[Xhxa1zc1W@D! ϭA MW8 ˲StȫXZYa*A&1ʾ|,way.F:w oA09vvˣ7`1i;-~g@>B~ ^)3D+>zQio.# D]&>x=xЛ>%-dq5{"NC 7!>7 MnXNȿv[1|KN"e^oT>j>!i3G1ATO'6MR|I U_6`NהWcqU(14W&ɤp]t1BQ+%$b.]1 )0@MA*0d D6A@#b ,Ooj~E0+>s8o] %B҇*^C/`bg'?.NϼH |çWO/Z,vɯo,ZggOo4zTz_R?H,Wd#ęśzQvt&L3o2=W7Gf'7{hXIpf߯~W?Ck_?)BQ yPgWsٝvY{6w9OKXŝ~שyO~.ػQbv!hiROW2tQ7~%*Bٜc܃s{[6aAz^9c\+b Vx"IךE7pLJ5DoOg"=:Q='A-V}\XȒ'*Ο]ϟ9w/2sTQ;s#6 A`sC;0z9aD<Ȍ4zKBþwG م%)2~ -3=!hzHM?vn}#GoOVIydyYy.<)3A@}0bp w! 9\y:Q3bƮ-Tgc} D^B&Ι\SW%ף+&/;o'/}3`Q!&12y'/| a^_9eMA/y^;ܠV(Ok^[Dk$ј[6y>h F=GcKBɀp4&ƺ#^,DeQyvraMq :;J%$fNЇׯqԨ=[wQK<+:f| #Z6SqJQ32'?CkهԄ۔G¼&{HRzҒT7!8 }钌ьl< YM>]*Un H#MS2, |.wy~YFRs`yHSѶ;L~QyM N!.߱ 9ko6;x}kR,}gm/հ排 !gx%#l-~W[Mxt %Ɛguw-(Yhऱ^{x/=a2;2Ŝc|PrA);b*NA@n)#ij9B4=6[6!לi1Av؇:o'kyBɯuYէyS-)hk j?h?Fònϓ6*_OE'zy˚t;䪶O~1+a;ߣnGoWĞ׫EG_]ѢGj.t:XuV:x:VkG 4L.;E>J3??\y>CBk@y rO+JuO%1, / D+o TKD^T?19薕(4?_[Qk  bK*嚀4.UT+,:*ߋSŔiYe],T\ E~[mD<:!Zgj#jO |?# ;鹝·χS]=E&Y˴#f!>Ͱ腐}MIkIqʢ%ىlA.haSR3 MGҭb֖ŒCӠmB=1[8m[4Ћ#@[Z)J5SNs.uEUtV++7(z'q"ӒK0MdIL^ƤSA-@1Œu0֎id/JNybG *c+Q{:'\144zxH*`S58AB y0E|C{ȷ`W](u *1関j`Tdʂ:۴ ){_JR9*Yz)U&KQuݚH#tEmH=]ܧwItceM3;.."PY&c1l'Wоn54UQʦ'x_*+^DES-l`-3@,<lۥ&jE ! Zph,M;`xOq {T~e~`E8V=(Z2ٵR0pknPm)0)a-l2 )0[R6 b5(,Ǯېti-$+5jjCY?NQ "QM=xوD0ޒɅ٠vo$@6-kOl/3-0lzޚMn34Oӭ&M Ц#MDji}b--BIY1zENN*ML:T@\˅F$5d(뼅b<e\p'3GRxZh8{{>/Ag62fgpK* luKҋthmܒqK/$8`}c"fޯеX!QVꦲ+lDnlj2Pg+yYk,d-9.؝~&D/ۋC|5t\jtx\ &ɟAY7ޑqO&ԚEHJp7&,dY/<Qpa!Kֺ%֒Jl- S. )`{Ȱy- \OS|No{>8q_n/K{Ǚ/n="`QiVJWo@V[u+%<ś#P/tR+eV?4^2)X2:Bыԩ Agۖi(JEBn$ohոBdr˥\1}BXks%lΗTtCN3f#8!ajjTq8V7 !7>zք͠aـn0=b}[4rO"t&C G sq7osȩCNrS9v͑YFf@#3O"*䫀YE>\>BV#Iyn"shB!wK EXf7Lk˺ 3Z!^ѱR*\Z|9;#(Fqb)TJѶz$U%2b0y\x,s|yC<xhy VYȸ3~9;O^nfg;Ѻ<[[ؑp?:MnEpm|LnKVQ_{Gahsd&ZOԝnkoߝL!ag`Lԃ˸t4+0ymޚKIM!FUqSĄ `BƅݝKɨyA=\FppV6*Q~Yg *7 \Rl LJIE|u1轠[]c*`[q hwf= ^=XZ-BI @YW X#W MUˆeUn|gopIȇ}r!#4e|bKCᤸ]78 )]Ƚd+8mwʸZ88}y9xQ+ai\ًeIQUta&- -vc&fc5e>jX+|#ĥ-tM[J :*AݒC}e 4j'ªʾR%ߟohUO$_4 .䍶Ӊ'jXU]|m;[W^gh㬳ۜ<*+ے|lʒnu+-U&UKrjAڨ\}sOnm@NK=PKmAI?]|<ۛ$ˆ+LQ AYZ3^ߖQ`L9!@Pmi\ɕ \0(GhdbMq:#bD6uNחlpǬa_;I7ZT9ջ}$5u&^L*$&-VmfglE(ϓ*c-!T:4 `jICEQ^{y0 DSzt3$Ymh2 p[!&2RZJ" e&u_xG(ݯ|g~…E}[֢OS:}N~+O?,g;&vVۗޜOO;|ɣIobrH ⽓7@ʤc ֣sv!}%eQ-6*3vS <5gN=f1dl7F0 ;3u&D[$f_3Z+SeBhgj7RqpZ1~7?osGWg-F$Y;NgJ%,SGIK3,$y~ 55B9^tQI. 1si] nsB.8\@] b _hvbE#r=0т&4t3AU3J⸹l֍0PniʳEÚ쨳9KY]m@iǨ&>B@/ Udze\a-Qb2mF05)eDSv$:$kv*q?jf=)ՑK!s a8rhKЈQ A_$o! ?Db[ghG@d} Im\$h7Zs'pҰqƣO.$-.WJ>a듔tRv}߼`Bol _vn+Qv1s5ov7S;\~JΫBv}9Rw}IgBnv?p@K77WWZ9/%ٕo6'v5CLO8HrVJ]?bZץv| Vm9Tps«k)Q{7~v?$_ Sv7ʖv/ŕw g\\ݏA0.WoTrsmrv+4ڕ.`}J.áoƮϝֺuJp})Gvdnv_9*.ۭarYuzv[ܵRa7y>o^y,1ZR\G/A<AemyQ{A ] >hՓ["0=_Pyny0_IWyʣIVy# 'Tp9K)p`JhGOTshv?pg[Q;hpA7GJ(+٩Q<%jo)y:kYJVjw .B^ae?xw䏓W,A;5}4 S9A=}2ZGJtG k pRK1F Q!a֓==@D+];.3p%5˟ouVXF%eS =guClEӰ`Z+b1"+$UޓakbtZb+q]lfT>9.x+9Xu/ b|콙٦I(q.+A U _"ݲ"ł<".0dLX"Yd`7_$m-C0!@Ar2:!Β͜A!0xXje?%VH8zC&Q+;KRmB>f-~_|{{nF4҈b - h{!CIL'{,M8rOj 214̬Q Yn)6|B2Sq!(|RL;/~Q geQ|~.?w2d_i>{vQ=|^|bf?|~YL/`M}I^W/dՋ[we_m`.mk7=|\~wi_|3̫N`-R4ZCor΁hB|!im CR $9!%ypNjW|U#~|cjnɐL5{IA*` h xG#>h )28.ޑ V&(29+UtDQ ji{𷈰#Zś\_ݦC'7b (▝L֖ ev"bn6C䃄g bjG)D95[2mJdZs "ˢIb; :x~)rq-Lô 3@ eBm@ +yZ$Ƅ+ѥt/'5t[{\:` 쏺(uh@N?0G&WηW#rlYo6ajfZkHD0g}E}d)МD ~_mçxR{?Wyۓ5Az GR"dU" pJV"mwjB+ąM}`cWM_;˵o EUFs۫wq# fU*jcdIac&7:hDnRĎjD%Y޳mlW}P@>om 3[l&- -Q2m"84̜s|F `x]wY% y5Q\HwhBr"٪\l#Ŏ$`*AU!"`kȒw S%VItRFE2[(ZPa@U H7 tԦ%7)(aI)kO)*`(c(]NY$+Yc;)XKh N&r@͔VMAzRBQ2T!,.5AV!M\K@S ߟ03Ұ:ďҌt2p_, 3ЗC@z;&8A Yj#bG(RI3l5iTx[-TG릢ZO,(6_fc(¥qZGୟwM MA BWu/toQ[5 %}E&8yqqH׋yeNeԗ:?#"" $l"NB:ľYZ`q+zReiFwYq_LmMȗk^*ءOrں=(ۍJr9ɔ_|1䆼 !>jh헪8h/ϲ'b?Ɲ=Xma/9 k(Faz.OsYB1âEj߿#Ic7V!T[i=gI3Slp{]@ Si8RZ3(.1(X#S)̞MS29k]R#@p*"A P15і9`ۍUUy2ңKvWBQم͉Er+i kToF&pBXu🛃f%ϟ(ol~3/>.3k&w,2tǫ}ֳ4J7UV->[t:ϴAf^|'Apz_M˻(_2QҚBɌQ%k'4 Z@qsLh$<;#f kr,1TU: H ѦöNݼ qfg)9X+a_ÉZ&{ߎuQ3ƴfR(I 5b}L|nfqrgxMOGSFQT]+DG.Ey-0I= AaQgww T-xyqQ=,.AR_` agfƝo36r:o4ORw?ahe0 -+*]ƃhWAH)I#C$%2 ~8 ? 5<=m B`.hm@kw$%d%~HǛ̅G]v<,G͊ V~]2AjRW'~/]'~PO7csN8ΰ^.wdjeMT\8-- ^H;{U&bk!QZ+0kϾ_)HBYd}&Lg}&L S QZ4"Α4!&8$NAYT &Nc+T5&t/T3A@# t}fnF_3KF!Bm"KY,Ԅ,2Ǯ,.*>/蒘@ }%Dʒ?%1}ԁS1hrMxk~BeW%͐Fҷ|N1d)-`E>U3V۫žjO(9q9cRر ʱ&6 ŷ8ps#(vrcS?~['EigjS2 {q}ir^K8uAm%ZRUoWגGy-*EFQxf>gK\1j9mY&R TT@1X/܁b +ﭤJbSS3G:[_Z:bU'ws<1~r*:} LJsPy\KRÏ~w~Ox'>ZkBMYx%p^ eC糧Mrvhspd:o!V/cR)(O6 3ڥ!B)b`J4qkE/$f|FсCZTcB.jQZTAсm}`'f"!HR\u̟#ZZ6 q0pX% <9xc_d{Z X<]MnX0K Ju٘Q@ &dž 24I-ArD*ÜFQOGIyT vPhQy)p4 $0Mk- DR -Ejvc]MB,,Fa'61NUIrxG'>4.tVp,9|54\5foNPӈh(!(P*# bWDÕ{k.k/S G; xqz}z^1 X*)qc'XB4C:`!<&%TScTqOrpDAQMCMtd&~Q෢%U .0u(8/~*.),j||ƽI/ؼ`E vwv^2xvڗn۝Nߢˑ,Y<,_ɯŪbI:sקK}RV hE/,VQnRJ0pkx||uAJ g](cČUY{h&?. ˠax*ٺۆ'E.cV]?`1ũɲ;TPܠ1Uq{U"F%⌡2\aKiGآPGr9F8:zåݟM6m"}ai7N_ %܃ Rޟe2̊8iE"wޤ)R`gp2>݌?\&W)q}Ã'4%88 ߖzĻ7?2xA1鮥_[<Dǖ(J8RB̟͜Mr){ z%&WF"BB8qwwhjWQonEZ"gLGO 6aBmYӋkwwwӯeZ-^1z6yp,70^_@mig6lWО{|>׬{»c_P *}%ij]x+b˗o/^^C`tV{{:?ULͿJg8փ$wvr$mlfVmm6{r2\e R|˟޹w5q>UJ&{]E.%bg+A&"l-xթŵ.[9H n+Z)((HIwM΄ ;" 33)eX7g, {2yR(yCALj֬ۺ?c]5Ѡ}C^%mx3AՋͽ1vuΖcb֛]e&L]A첍_NQFD3 ׷=v8AZ>@MJ Ԥ@MJ Ԥ_5Ws&\R H9Z0v2b6hoC}=BcCmo̊I:T/vDstebix_I2g1C2H{L$j鏳a @9>gA9xgyي|1JCJ)TB RBۻHvP*ʯ}|=0=@agAëjO-g bl>43;*nE[sH2gc6Ne;c"*)*|U yp/Q{H ەG~?Q&V gE.(i="*C8ȑ;Xo8IZ{_ Tegi꺷:Zu GtQr3ʆuD)iz6z}>rb5DS1Dn)+iv )1:c|~6'CrO[yt a,,7?~3Ŗy8e5|| v6haiHC#J(9c@@8X/:e$MZ L )Rݺ) M~!|Z;܂")b\ia.F,hy0IlΌgv{Vw簝Hۙ" us6-B8D_F^Eem}Y8TaIo+5MXj!v67w=^o|'KKnϟPM&_Zqz`d ʖ"Lﲍd!m+8_g]^A&ȚcEzKK׌U?_N9ggYQJ']܇'L] ;׻2E&n `$l49 ɖpNߺ ڢmq87l ٲ-n-[C*P]Cr/6> {Wb 5vs"G1t]H~2xxm vֵ8OV?g+}ړַ&P={uM}}#YOMY'ދߓlg:h2Bet4[\SpΫo;]ryy;ā mfwYYs'. nEeG@1ъ_І"%ɤeB)Gœ׉Z8!/' Db93:*/Iz! ZjP@݋IH9)DB9)D>f_ gdJ.Y7AAAhE'n*ŪEP[_4v,mqUM =*1"۾7c:tRҡt褟[Xqm&>;MKp.ŀ6't։.&(,j Pmm|e㽄GIц8yhMڼh+/ΏX`h3p+Gk 2"XcN a;5Y 1 eP( Y9!{4~h^[w0?TpxXƢ1<9mMκ)}:gGJTό sI:k2G40E QK2@+_Q굻=ݵv֮w[hB6)hB6ӣ$a`H2DH2%uP^XBUȿ2[j[_Te8c2F-w)ڌncE17R&hˍ@Snd@|וqZLޠJQ4L/b i|4d$LKj&@~w• wX`"jsf$ȵG)sJb5D@ggW2*Sp*`QDUyH8/B2ɨ}JrW\eM&:+gA2Hx AVfT Mvԩ͂:a!1͌)t29K+Xrkɤ􋨍Y)T c1B(vI3;uZx-hQ t봨?[ \"/L{"҂IˠR$ZZ?hFTQ"g#4lC( #+dz\Oȼ1\I{Q&aܠ&3[(hU #H4'yrͲ@4mx*ο3<5Ms`XBiCHs 2&9-EXe1@om`F#ck+2 ʊWCsA kb1L{̉T!;M#rl%yN?NDyO`S W%am6Yaz`Fءe\\7lf(%yS*pK)~}5cnr OKIlow3Z#֬'0ϓ9T'o]/'6ؚY׬L@n~ ۺrv׋7~v/K}'wt]~'[d\#%0bq𾑅a*)dC^G ZmDxM0ت"[.9Pc'`a٣ѵ N(=y4[ZlNݘa\V^(Y c]A[dA@==lYyg J}0R!z~ӻqoWH'npCmk?bǰCdM!*?8ܩF)ȏ#(7 1ПV  mh2F)Gy+W?: wZap.d'Ah%5i<xMkسyk_-++%Y%E_Z0j9cdZBvhYY]9 ؽ:2LyJ̣r$=D%}B$\ 6EeH(ʵ#,xxw A6Rï2T vEf5W 󸴌^xZ ˅% &"ԋc[3Aƺf*^5.w7Ƙl}w)S]Hɬ-^H Wzi&_|$jY3K,No-G i&uH*:cا~kJGr|:8.=p͜9fkNLp&pPknN$8zֆȧX-0ò):er|ܠy>]^  1 u8Gɥ#G5,թde3ELF%p$KyYC m8|傏Sݺ?0JoO0,vU-OٕMH;(2gZ㸑",0.l/p&H6fk,G'߯85-~3gYU?֋*t̓!4`"l蕼. h͗RuVD8 tyEa]T6ZyꦖZ)4R! -j)n4DhXc1XcjII#aWW׶X>+ffhg@Yok]M  UNJj8PE {F,֌'%$è88  ڈ(=BBiyAk8U W#_U(JXc!Vh{(<@( Zq^ֵZp/RȊ"ڎdQτ h!e'mZ Q17K`D86-)74? ބCSN+X'LNL=J6BQ(wYhP&Ƞ+`dbVc(Ч^bK&Y8d%8Fոb@_ݓMOѿ;o^Pts{.//ޝw}ǜǍ|r͟5 >.ݟXOGh&Qv[E1)vJa\-_l*XK斳]T(CtO?C7]/ |ԁ8L&l_R6:2v w,!2&spǕ yFJ۰ vṴ;@0Qq%x0al0*!Nه9%BeNŊi;`g,Ige9 ɆݎϔdwJuv&6{,D TS>%Y@E ŠE ,'!YS` }0Aގ]@虃}]/ m=CNBx j/pHwIˌA8Сkq!AE<8\RvR@!A6vz$ I+S/%r`/gRGأ^ g(LV.JTF-'`qaA@˶ xeH.[N~8EEF'Kdh/w/ v0Z`ѵO^ô4!o(6w;v4l}f ZsiF@iϡ()$H˗O2Y$gԏR3g?+Xeg4J.I欤cN;"Q̄8d9D鍟 o#Mkg#[:cvoܵU5E:Kd uw'65>XV9J$ß\Y)%Ljk>6|f2מX/cjG(yzPCsqK?mY^:[}r^|{1Dҋ i%sۣOcrMMfg s%4h9Khţ`(K).75%o-B i ![T R` k)h RZtsEP'ZiŤ SCrF A<[T(x :qb.^Sn{O7߾I莚7\8xA.͢t~d 1[t1pG -x|û|};9k.ߺL'CKX]<;+!5`Kd|S[b2(Ax4vʧ̲oXKc]^G׆wwϵSWis7g){`u]Ooz}`^&tl"܎sV56ruuA$6YO=p(V{>|b;0nG&}oë>xOsoG|gb2F'#QdGфAx>niǡ>|^(nv/`~Vȋx dHk`PM-:5s,HrQbWm_QѪ'!!?e^myΥCi ǛpI34Z}-{C ac5#P[V޺b5M]{WY[BjW[On,c_*`Պ]6dcIwJ|2ݝ؉GJqg4}g;Z/)Û|˯H; nl A # !GA45HT_EvI1V5 `d~7uGdw5憔{]E 9ݟףpeD>)A, zEUA;fٿG-25ݯm&h^ǿ?socwuy=7L]Kh$|-=zK}ݰ5{AVʃ蔾#G-}E5RInc +hZGqj%[)RSda ׳ibBS!:Sn-ǝX8A Y܁b)Irj>~XZlb5)2=8f#՘&1lČe5y*!*/<`[Jr.s%*o٥ؽ2$jm;WX6*kKVEB<&fu&erCL@"C!v\|vשȎWx_9OVμV͋:۾,&m=x΄j%Ue%+ Uey,0Y|\G4P<jf&X>cÕLU*@L捨47:(\J5M!:O%0RW͵0_Z+oM-g4M\FDgDZMc qrˆ4.on@.7P׵ 5g`d艧4 9CBBu%$M6êU ?`նճ|3R@#lAJ}31CS\;nZ-JywhϦ݊ Mnc +hn3zHv ڭ))}Gv[2Fٴ[1m |)msre]5|rd!iⰛsit \>M]huy w<ܜw8sêݏWF5_~÷}dfbHV;}4cnVy_)Vv+A tJߑ zgnńj1CtS2SdҌ|5]*GPoJy(?^^oh$wG &QZ0>8w^ *\!if}~ yno/]}HIm`s\QQܯ@Ls쀞 $PC}KhH-۱BDXS,,VfXf'C,-k7X:OiAlCd-%CHfp:DҺe2H2L VKa"HT\q-!QT{^Qw F%J,sKx\8R/sK||22v2j]]#Ǖ#ܪ2jhű>)R)^S)MX*䰤5u͒}{[2ed&{]8ŵj,*#l #œ(J]>$,z礳4ej!'b=--K+y׏/uV q)CÝIs!?mI+*pw[cx?TK%L\d-H.fR߯A(YH!ZQ"AFh|`üU<5^8>8<>ŨD1Y]x%gKe|go6D_\&_8:~`l8HkCpBP}"Mwˈzdwf rA@jYB  t&)z}B3KGTXݺ_:F p՟|[*ݭS[tNnJx,c'%}$! % 3sF""ĸ oeH ڋ(T}߹v4!) &G@; BƱIt#ɽ&&_ʣ> g&lNt<(d(hXDWZXF;`(n5UX 4xja0Z.\ƺO=S/-oeD(~"0VR,ksJ*_+.F-MS¥U1Djs6.g.Y 7/˗?/t}Bb4d`#_d_AA0Ys[qy7Gdz!UThS 9b"TV6ҭ+814gCh ͥ8a?'A0\^|@wK+QSxΦpTϦRٯkyU-]p&̗lj.2r9V#tYm(QyNlSHVd5ƈ J8jnyBQ4=)5Rk+{aKP9nYIŴf|5~a:G,a "<__w; & (nfLW]{W~}fhE^ag iFp஺9(D ~pVǤT+kѝg_RTHjiuALqݚYr5R(/oo1*%x/Sk=$8YmJ!ƠJn*z`{0^Ǵ Xxz~]Ӊ#Rnܹw~q9i}+eYl;Y2%Wݿ$//1=qU {w_80NL:wi#gVݴleSƑ/WePF+W{n-L?u"{bP(C WZ=ZtYt6LG2&Ă (Дp=Qa`< ~4(dB?Bfgϵ@ )HZX4, NUDoVs=LԹ_ÛDAU1R4>e"op?WGVR:J ;z n}حQEְ B }Mڲ?3-ibq0ؤ<0q#}3K4A2z߹oSlf*폒֋Wx/0?`&}W! @AkJ/kX=|U1tu'sܹ;Z-r<'5wj37A?7՜$ϋw56-e ny:`T I^P|o:9t;*e֜wϚөALoE(m"b(K 󃲓rurNPqOw d8W%ª =Hfzq˰v]^<kT(HSjڀ6ǀYek y\k,zY b5BC=ߧhݍQ Qp[mwߜ_.C[Qzo72Q9l?x|k<9s:#uiZpt|\BM#mw1w׬y%nhd`\ ߨPH3,_cOs0pZjiE!^1+}4SG'EuuEY7sž%Kkʷ2”褄02D!..}cᵚ(T #XT4̰ rzMX8֚LZ_oZd5_RN~H-\;.8$CseMOL 4lRuO{i8-̽nL/}ؾ+<iz zi"MۦiNڹ>Dag1ζg+۝k\\yg sb4;Wуucq],s=M*2&Ÿ_5v*v 9~l3[GIZzgjAѸ9:p By&ݨOkgr0ԠLQ~5<8sgT"k~&(wIvfŨ:Z=S`T^s1ʙI\#h0m>)V~τclIr;O_ZQ"LOߨU*ej윌ZPju-jGuϞWFz:)K:t-9*:)ݱN^UT!V[4d:18J  ZCPmvwsMkcd/uH2C68ٽeUdw5Os( ^8LfRBs~EXyjٴ?Mi0D'zN@(cuN{BFH7ԃ$1+ҵ) uy΂70ĆFPˮՆyT)ٴ3p s狻;l5zx>U?[zE5eTH\Y1M y u:xPˇ ܄VvA3^l4]w-EGPJ8";W?{\k*04$e.L0t[TN}R=4ٸy5MuH8zh#ǓwN;ppc~O}սl\ȹ8.j<hOipac՛7W1dk|;a<@AcB5gXpwPKWbVsY\e÷+foN}Ͽ R'f:wm^3{cemTXne4Da4aMy 5PUw-”ٯyi`@!`&B!傎 B$BKBHC b9W~Yw6 2d6Q8F}| JW~:@LJo4($}-tA(A5ScgQW.hjYl4 3$#B ("2p1H! xGEHPZ辍o =ӓ^G`CoU*9R-80BYiV+R %/iI>i8'%L*!a|hBTȔ-@$3P) ¡ LV0йޮ)'xU0V<0howbweC{½"d >؇SH뛏,mn5mcZvU qP;5;LrdVML9 d,f;.<ޚ@eAnKﴈˀDhCgT%;Mm2릱(_iI/YSvH~-t_ aBDo1ڶF1T`EUe9?հ bÞǭ{Xj ϋN{irJK={~qw2fA3J[ɟ#K=[m>3 `>SRD[L{ |D؆o'ءdd*]kR#Ҡ'}؏¬Ec&Ad0Lj'D4h'M !Zy1>g3ޟelrS;]K{+Z=.!r"zH9J8s X( G 8c~.HZ=Z!QfBk; <;[܁^1몷cA@hDjuG.0cJੜl#AqS1*9=Kω[UMw[FCbU[Yb&gZ+LP$2jr 2??ă.w|{N/%6LNt;67}LnSd XS&;1>/GSӉ Iof̰֤%]j4O߮ϯ|5ML~Ro!ڈBTe4QYvZ|0ŏ#06W?F6~`o/KOzK/V.^y3xȖ.0tVYӯ[ov 6Sx1g;;_*u7h?gK2Z`F{s8{zo}QD K4 TD@FRafX"$0ֺ ʹdS243փ R)S͚a=J$ϋ;A1dřD!37ML1GjB='SI1]ת8~0Ë-|;*_J :Z0NƱJ ~A@9=SEa%AШŠJl;Z e E#}ߓR?{㶑08`%a8nvXg4%ۓIJ^#RMJX( Rd\H iHCT"Ribb,S'^K@&T c)) e8{3Ik-S0 ľ.Ovõbt0~=0&yVvxأgl>YyŞyo۞nrNu‘8FJ|R7ΛŲ#ܝUsp(ڈ{O@) ۣ@c3[#gP,R>3S5 _:blB=Az:S0/^C`3:70È48n/MՆkSrkC~# ُ קZOؗ9YL,߬Η1`RMu)rڄ. 7*>EBoZ,^j> n°@5ՉX|ZE!XT5@56 N'C@DFpr9FR\*}Fp$ybebe, xQoa\58|E7nzr W;8X^bXVasje,h5[+21@J-ж6krATP&Exӫ.xlE@1,ʤTcR`P6۳IZ$jIUP yge{L͜H ii2ٓS1E‰uHa 8jkn-N82'vβOw~R=qzw#0֧KjoeyCv9<ʔl+gH<wvYC2}t^PrA:_kޯt2_{yP?>AawI/ _WWwLW<+Σ'2읥 BDIؙ}FlB\I|JmFPTsmS|-$3eG!Z|>Bɼڤ Z(:nBkRQ 'F!UR%pDJ-DVRd+w(C\o8ڎkXH3zC3,eH" چC!p&j@#,bN:hJ2 O-QS38XXX&E@ǓY:ÏlS.9./V$㯗D^nS%g)aATIbH%nLI؉DqWyS +r&Hh-RPz\EzQBSf=0K*nsce 1q /=XX\˲7XVUglZ]c磼raqTwL+(%tUADW;jM 9^\tf NrO5ZK5Bx%^0ߔ5 Dȳ$FJOJ8a% Ug#"FUoQV[JߟK9CT`Vnu27egK]y8n2%E&yWW)_1lo09$J X_'a6Pi~P6|^)f$2a~Sь~-']W,gv4W(oY7AՉ}Gv,TYhFZ64W"Ol[7Vi{FAՉ}Gv83TYhFZ64W(s֭Z9-:&(@'nьֵnmh7Q:%8bbPGubĺ1o 2wmݢkАo\EtJҋ_ ٴnȻbPKu"Ⱥo} d*Nhmւ|*zZ*/߽9EA(Ju9 6qN̛7u?&Ӈ\Fyc/R`pо,_[t ^(]o{$OG@#HB:87qЭvmvsp8NL{yp=9VM)*o}?4M%1'kK1#Lpwei~yC#x"o'y#Yorj_窍Gư'oB_j *)_Qws\>ƣ'\0}@@eU|Q߆!*ݎ9x)W+DCVk̽qcBP6^cQJP.ONjuSˀKpKjy6\ S~\\KnVE %2LI.I.)ZI-Cyk_3&,󼮠@@4ʃ*vൟeomk_z*%ZP ZU*ZtjO֦R z*JwB՛Lq! m1O=3YE,l]UҊ%]nG}RhɌ^C#_V`+󉱀2//{ Ή Bdˏ};v~ч[ 5ށ5t8){,gJY*Nw[:IFẍ́GrC+pRnE_a vڜW ClLEP? f所aHk`ɎTaBQ+e͊fڂgQN kN )'\ pYV)T/ߐ}:s]A5~Fkv<+ZZ88JIgn3%+sʏ|CZQtQZk鲳O|c٧O5϶Pݛa#5 uBXI8& ;Œr4Upw1܋ r5IS;S1I^'M&rZ$:Ɖ\+ %u̥2E(>Ş7|VJʼnvB䂩HXꅕ)v\(OO-F1.TzI02>̵45ri"`PGca6T "H $ ѹR~ڧK!J"fN1$Ny@gwO'$ gd nzâI,)7a#%c6cٚgP0v:rӼs +0"r)5 x0 ZwxgcfK.iIzf[1H!pa 㼹qx*":S2)D\'but>g4_#CdoPkR <"h& =1尤0{Zs& +"#G,,8\Z ]P!vKcQVY=f4{xZy3} YO'^'AJG?&?,֔v:O hB%֘-K&'g30Mp !Ҁj~rw@gQU3Ŏt }>Doq@e2JޕWSbP|gu0f,sLt'R'2`".9l ~jAatyӑwS%W7cЬ81k"߼ٽL?u]g \EXh4g#<^V};3p&&zWӢt7 aR9p<88W8n \;Hto;kZLEG" ]%dG^jW~X*;J`04aTgjвUU:IN*&(RG>OPҲ `藜!:Wg.lY7o|ETGP'ܘ]Ļnejx*XvUݚʧ]gwwH9  Qd=Lf Z)wdڣX"$uRo<F+Em dO>a6S7k+BFpN<' ,GTβ B1F:0wR6ڌBp͜&.D^nJOxBB.CC, *M3`FqV%<ڽM.}ߧf>-'O^u#?sBcH2lVP֟M>,0D %^1&O{ 5SE.1ϩb}AD߀mN%uـ)yX}5 F#jR*D;HF.gD+0#6n8@2Imn~dTuz>ߎg(L>s}Z_w rEN;p_I-m޹9_t 1rea%[{zU24;x;f1OkQVC2|,cV>(&XGIn<@Anj3 {4 7Y&b}-\>ϺuGd<4v$ #"($̛/k5aA)IyVgsU}&& Jmd*.g~>IAFb6*stOŊ`q\xhOdT3#/ XJ@pّz2@>||&rX޳m,W:ũ}?CEq&h~-r+ÑwzQi\.hc;3;;S MB8Vc)wʼT$ꣅ1d +hZhń}^<LjhB`Pe.LY0z +IP q!g$O*ک RH=ngYip0C-,_,T F5&:f+&rW%iR*:\@} @jg_08 s9ExZ˚'wz$~[t+="a@$v/}x;6bه;ѡB۩vN^1S#T{='!xiT%k@ ش%1)8en]Ϊ.%+YM?)jq[`i =KMoEv shD2&n7EҷĽRIgZ:ʂķTb UϏ|h9əU{`X2qq Xɸdٯ9XS2%Pn @Is 3Q RHJ6!/UMR=y[O:ja#"z(,CKe_n6u74 &;kgqxx( @n+I&;xkbӖ[n@wOu5b͗ DHsa\8z|VpUxG={WFD#)49pv Ba75+Ṟ&K.*SI{<4D΀}rm"bnQc T ;+C$Bmˤs9:P%C'=0\≠n;<#K~v|,1ÍvRW/U\S5ɽcD{x !>ׇqlwR=rդr)8kB,}<3_̍ܩ@~.ʛ$tODY=HKq`D7!fo֯}EUBZ j}E%t7pCW1Dx{~{}1Fj~B/πgXOIlEB|Jxr܋0տ 0bkz1x0 {;W<_q8u3+FR\lOo~`]T<^O+澑X:hv(+OjywC!;ŷû,`:Zr͇N7!>8[G06A? \YvaA6^7CDOdida~bч""'(g%2"G,JhCLd8gY̖ a0 4P׿5 TǛ[^i⿛wTng|>!?@H'%}fMcM/s(7BϖB{]g/؅ Hz>3*Ut(5Q9&!Ja%pbz?;+GˌQ%PBԜ!lhdR'!]4D #3.|EQ nnunڅC6F ܈ IZXosCdf#)-E(\'L am);2P³':$LZ-)*% ]+׉nCvPR3f<ٱj!R3 [Zג(gR" dLT7gJY-bL #K'Scd[Nɮpdm96L`rpO@R~^jH&A8P 6ĉ XiHf`.KM(̸6Z' Z%[ź VY0KX6d*C%|qNH`0.[V* Xʤ RĢ;?O]&&$ :Z\;vznXPIa*+tQY#DA[t7 zM?ztD;[rYkfX0Y5՜'1wMb6*-K P'17t,ypKa^FŜrUc@;`-{ΦI_~${ﰌfl}Xxg(~Y^\ n:|O҅e%xpDcB!׎ENJ)!d6|;Z^0CÕΗO?Fs7O%&Ԃc_!Vb@{ʹ#|_qnM59RZl{Gfgkz||9>گ'.g<+U[vO=31{LXK +#%'gX3SWy)m&%iZ&XO~2nx)Gkv#;19@whHl{^N/X<{a^9ӝ]l /p%4yv^ V`ˋ_>ÞTJI:ճ*uJpLE)q7 Q ТIq8 wX\LVsC93=LKbn@7>2Z뭳KU+fZR߼0R Nj| \|;XQÝAL-݋INcO9[1F9-]vSҬI}iEo5ѫVzg~~n^e zgHZ{3o4jEG8 ~F8?O{یTc+F]n˷~`ܵ+б#Wd?u6g1M~>¹Oyk7p@mq<$Z#bݣF7[ktC Hє,ϭzǒVܪurKQާn5I6ѭ:5ݪqyrV$nuYɹ{`< U樐OM.SpTTrm.=.)$ŤOm]2U˻{*$hWۣ6:4OxI4OerVݽ'`a- }h!Ҍ:`Ěgt_c cYR9Ps@(R Y#zZVYR1g?Ey<'COӼ**$^G`^W BeW%\ezՉJ-O2jbD lAΉ¨!tHZ\ᱨ4qc^Qhl<ՐN& `L:Z}I|ˋe~(𸰬=KL7PS? | |$fPl%rHp{3d1Dl[B%o b^4N=K>s3 9< g¦ "X(X^:4Jne:.*tP>܌0%{i7&5XuHFU~У67=՛.JJ zz_`i'@A Jim̝U[qO~^p>oG~lj.~_CrrF p3PbcA#Q* >6vJ0$RH@Q^hvlخwޜp~/n~Ն]6?:WutarnMH5HlNTpjLka$"A2LT%m CQ_E#sJT9Sd92SMDdJIk&MݤeAVM\ JqZ2P>X*q¿6"s) X}7|iS$BJv=sO)`eQY8ใEHhC9, <( jU*˰#)m ٶn})fYJR`cwWm,KSCnsČ#dcilabBWX|Fn63^7^ XMa:1vAa-}x| 2t RKpBqA@/m4$ 9"WڊR8b\&X\!2QTv Rǡl-8 tY]2G:b$Hkue TDZc+0פ@kĕpHJ ud9("x;TI)r\PQF4 $-p C43b=TJvY*%)h%}ql঴HBl#IFhb(&\Z8KS,)8 ' aŕwHR FA9)@19Ha|`R %:8Ĺo[2fTNU39tn{V%Ԙ1 -hHd@JHaJJdP e b`4iiR 0Չ4@AZ ,CX@Tap,LXSXCZ"F(WUd6Q}"^+;ibX"+Іч_rž2InTE0nNA3$Zm~\c^fI.wdn$ x!溜Ľ؝aEh<?73nd._Zw'(¢|2=g(v)`}髋pRsMu -nu > ]41 4@OǦR$c+~fbW|{m/mqњ/~nCpm$Mó׍]۱4i~$%;e8L3p8>Pv;~)GQR%=u8(e&vf.#N~MMr{V^0$]T7=Lg^x-tYu^^#.WXDڃKEf56?rvZ<4<9GZk?*泛T[mΔX=Qݪc"'1cvb0%+j1bMp PK?ʘ~~I,3 J=$P>{"0RNS0> n1tW𤋮.uk!lyl9{ զ 4M FMϼ]}y odh6=G29Iy~LÐRSCw\R{%s>\y~*jAʇѽilj j gWA㎗L&ao2IϋukPg,R_uTcE;V1~M>,?ۃxj7G#LiX" C4t/!>D (c& & \ID4" S,!D!L(Hǯj Rd$5xO'w@p:=oly/zBe4|V?{~KUCN$7l6 TG}luHì ]œݡޅA)m NO8P H=NAyj+3=xO'$)ě7O}k Zyl-?g/_}yhѷVc6\tn^ݾ~͜E]l2C,zy?5uwsd:՘'߃ X$yo/oLhj뽒Rkbtn^(G7JI^pjy wu)Bi$byvBQAMAW=<N(5 V*B47N_<WP=q4|kff0x+8R{ɺ Fl^\iOɃTmQ))Y%jETОg܃S5/e_f_ut|VTСZs\Wkw.///ޙ_Y9~ j(.]vw /p1.fl 0$X+o5\@ 0*/0շT,݆qq1] օZeZGJRNv:hv̹B+{<071" Ю^W䧁_y2loi5cɇP|d#l0SցCs̺|5(UyOԄ"?&J;}RD^nqD@G8MS$%NC!1C#@dۄʼI45)$JTyRנ jp6=X|.EZvc]oV  ">y)*aHBC (fx4yL)(A%c"X(SPUa9f qΠsf'X'Ty{5 #ȌsKLQ=^MWAF_Ilg>}xL^hųIP 0O/^Ch@EQhpno :Ą~<& uG| ?/:jхp{53;W=CM,N]l|=U{ شpJ)Y=^1q%(BHZɽ /m\qOiIq@rtWj  [6vT=ߏGJ`WJEJzYTe$ijJ+@:f:}x߿>R?n1[%!A:X1mp[srW(iK.VI2G;|\q "3?:oqd5{ŵbl .2":CCq]@ P@D!$Iat^+5n 1 W&F (t):켟3tIw嚞ΠhUyW{ 2J[ eO/XTG]-蓿C}A`b*/砜)+,13#XD Z^Oġ !2E2bG {}_l)U\=كXPJv_ Bl )QH| R̐t:T퇜z.]΃R 8t3(1rFL.GK,?L$jHTg=AdӬAKާezFQO+q<^tF\P;Y~٢5~_yfN?'W J̬.Rp*q]"OPS;;E跴quHJ󺭽SPo4lMboܾF1 [HHۜ;Q"Z9[xT=.}#ǃTF^+qJ,SPH 2H$ `1)#x P))ļѓ_ƪ6E:]>y}J7MOPS-F[=D0_Y&gO )}cdoEADٲRYh *DC|ƥHM]f-|ajmcӝ#őG|ĉ;#L X2Z:UGF?<k#zɬdZʢ|P3<8ٵc"q7V#oǣ"MZ -kjL&}^y E9WYNZ"es#eߥ S 09mvۯrVa-KJj<ƊVW_- V˯Ys{ B/7s\j gKǺG?ƀcpX#%m5g5W1(&ϴrP)13b?{4l#x[?we >{H󛍽d|.gN]~e ~҃bF#_:2@02?f"f5J!CuV 1 ~L#DYac&EJFPnNxJ $ %HPǘ Le PP0crID ?iD\erJqKh BV0kFQXFT˭ej@;Baj]D%ڦߘ}AIcNr>J$k^䘸Ajӊ̥~ϚB=s19 9g+n.ϑ $+x8Kƕqep9+Eeosfϙ{K#"I5=GA(O-Pv>U㝥OC7TlaZ,XSMϖ˯-:[%;4aSNQP3gj */:O%I8sDESa9vJ/vSʧjF8&vJVtS^ҁSꇤΈ@iVH<9fY.4IzFGa~#2?ػ6nlXl{qۛi_va1t4~II)id͈)4C;<Q0G**3RAW8pPB^(MUqIUҠZMtT`VpQT8M\A`![Pumy+ϔYnjKr .ݒP 'C2+ϧUjot$T@"-A(_%f B}Pg]%&B1 H9V^u3Pb@"N+K@Jɩ9!1XS-$3 TBfxlEPN%Mc* Y!iPvIF%48p*N gGssϴЬWiTk*@a" Xr[@Rh5e;O:%u'ɛޛH`鋖 tM=F>VN=FֻaS(C8:P 欒gHD U֬dC蒚ʃ:-(Sh% n쒚CB^&90L5qKzS' NNAONSe\mn59$䕋hLQ)"dW\OO!viB7ZYr6u=ebs_ZncBup]M!qR\ ښbaouYwvC8>p˛Ng;@+oއeeA̷%Wzw5rKr;;O,t$MD kq-ɫ&*W)$0x;6ĆSVRдhtG 8Kb+{ʙBYz50gdS5F؋O *@z4|iPP;x)pԒmg%z3$0a>RUK㙫'gkhAxw0#dfƍay띙`4 FvN fQ9k4< W{PLWډpܫa"QM%csEeB&?UyRjn{ނB.ބA ǷS? }7󛝾(w~Q-`ٟ,XgezM VQ VKh `3H#rf?:[l@ͧ߂?NNCl"rk$uU$d#F.)7#6yEbS:qs,:j zT•RHrék' q4kckKw@Iҝ:u*^ ޜ?]ޜz:GEO˛*թBU [5&QQ"+dIk 0EjjSn - t&s !q$6OoPV n uxy27H!ϒ[1Ib#nj庱9yq@KՊ [ra$BJ O0aC#*o{(i5UkVIڑ.Fo)&2I$Ϗ[05o!M2ңjjfvb6QM\;dpH#F'$ܫ͚u2!`%,6;Z~JKP5>}̟#$YK~/GXΗܥPxz)(n"E|6vq7AЗEpYnrj/qwu|(SRN6$/JN?v[s.-yؠMyQ6,,oB](g,5N.NbC0*`6#V*X}Z]Cc ª1^#LB[3 =e8jXj`bB>{c j[~ %vJ@2(y*9zhjh8i?X30efXJBm'X{47HA=)8J^luIJfT! fr=w`oZ),gXq!FfR(-3PћTY &\7Sԙ! t,TA,QSȯtᐱQSV43tp0%jJռDP.6#*PB7Q#vX*h5iʴV@XzF84I X=P[%\o̻t;0Y`mzq KVʼnfXEo$~I'$!Aî(Ak7`]hѬT֓u{{j6~rJbOb(D1]G;f58,Y$3DY-K̭0NTpz2Clݣ?KžS%ARoL>|1/Y칤QLYL,\ogs#ʜA}2eWOOq\X'x1]=|N=px;/E᫟ gzysrbiO?;{v;_4ɑ?.Ŀsn߯4}߹^1ǻ+|`ܖ_T=Z}hID.S\&x!wgaEE;7?§03Ss no],Zr %xv?jUk&M95嘟nJ)]r^劯6ǎ>|;^-?AwgjVw4ZĈP-}_6H8=9Zk?ْ̀I"`ۀHvWt(VO)8Q!qeY \to^_MbY( iZiL'3+U>t>j=DWJȗH*Ҏbq@ƚPF%rLk-(ղ0sb曙gOtk?f*R\]^f57]~gZ (> wey, +B^}t{;=;Űz-SR#4 Ep7D1J ۾kLkͿ歍6/ה2XM:v@k2eG L'[zBmD>*]l)L >P{ފzBz(U=H 0$'ڽ~a4Va Ur BN  )m`(_g{1lz1^>ro~+Bf˝p{zp"Y;}[܇VZrn  F9¢Hq+R;. H%Z d{"i%@kvq6A A; kk|i(;5]X᷂ږOwb4rGJ(\lXuU-m!D]ސ>_}7oz.`] \$M0Z(euRk ?ˡaVgJZU.`|_%6'zK^/7ၭGq0}z.o'mZm/D͡})]{o+”rPWkkeO|4XUGX6P2EBvu{ۯ? YϕG;:uG,2mx .*g-$_JF9q I4~"0~N'Q۩W~̼N2ɼjCf?DI(4^5&.ut 5(Zc'>z)!Dqn?q9`陋(JD3E0 3 Vq59T㡙ҎcB$qu+q, 38N=8oQ-h* R+1CC;5Аx8'W%'tt5( z5_4)}HMGGc\H ;intKᡮ֓ t0Gb*'\) %j!pe+ׄBZ T9J! X[ skjqVBB4BwlRҨlJΆïw^٥ x>Q} uf܌j>k(T•- oPNP\9L߮l"Ol[ٳ(2 Cwg-)t[CkXS@FJ6G%FaCOڏ%.%,]?wZidvN2.]+h+)v/àttc7lA~|m~m~yxPң.cem}PD=ʹJ-hh$w,TFQ+1P@ CPj\ݳ:Ugf06PTdz[B.y LzKVfQ[0-D+fMd|[;D{;nP5^_\~2 \}D&N +Pm=E|҉2%[ij}¾7yqjZyUki%Ux5edcΰl8(#Gw0zQADHN[!P}"XZ7ׯZ(=sKfΗz<,?˲j* \Q0* xthcCHqjS.A(7 XO;,!e9cr ?(pPƝrey=/URM)uϖ'23I(e_" bWE`bim}=,AR=VL)=ڧ {#J${$4GgZ*p%C ""duH`Ym&H,-5`5X$1 y]n urM nK8L n%ڃvq4JGcy~X =HSQ`5b+lt@Ź!Ua(I_\FLm=ft;+SH[gmU'r9~x*:UCg=t^KOCąǿl"bT[|$`86 }1$R;I^;(NB\{΍V9ƹ$C\QcFoUWaTFG5:cTITSxCPggZP봕Z#s )sbyNC5=;FTQ^pdkPnC*[a AJxrϽ }c<"/"'XȩYX^8Ӎȩ+cz2KE~~>Z{:ՐPhr ah-_n:}-0~ĤKԢGӠ#i=Pӄ BFB(pkX\6Np9feu9kEmn='-)HIZO_Tu )m;5 A\<ڠl6bb@Gtɻ_ 6x-[>L (vgPDLf 0/z؞js{b1e a _St8Tqٲ麊Ĕ)iɵOKG@DIn՜[D&E)+=AH gN`wwZW&qWƸJq)2PS+?ER,c T"%ϕi+Dq7abgv:-bco=]Y4Yp4:_oy? C?;k~09<~C ڰiDP>ZJ:]ًC6S-Ðoӭ{OJ#LOIl91_* q ,ּ_#&Yq,g'kĉvq1ւ`8n5sx'hN))S.ϵ[ѡD=ZM;6hba> |of =S26.2A3&Qi+$i1sl= mTMBXSR= Hsu-Bu\%uY)W``rusu[Gq4q)p2΀ G:4z@|(fWY7J>fcԂ RMSZ4UeG/8.kcۨ%BJ9m0SxM <'ai(Vyni-cS^iOdj) ="C7W[$Z)YϵˮWҫkўJ#:zUf0/y@!(?y ns֚pB#q_tqhMgyتL/&c1M uߣWB~Ё= ڕO<}6N|v!1O13]=?LydA]n\ކl[NӟJ8ge]Ņ,D),?+A8vYrM)nYcI0:C D=]ûo MH+nyDʃ:FxD]@'5DZ:$䕋hLit+C [sTb[m_iA ݻX*ۤ#][8Q4F+; iyX7k@ƥ>A{lV®Atpsgǧe$\ݫZ$CT"<@)e`26l"JXV&̚!Z5a3ċ•h4A\$GcNZ\ zK[ :(Ǽ.%\&Fhze0,nou(#FPNEf*S #>^a6*>Ӿ!`\ڞ=~|3eflT8\t/;'D{(z&⋃39" w +Q 3F{xZMY:]3nM *LK s!(P#h \rLvI%bFى^$ &I#w VC-jԪMâ=?F:7yj*8^0Vϵܪ'F t)evGxq_~csE=c| RG"{Hž;ώN$im+ ySz5]R&Nd]|?06٩ gz|%Xc7ԞOzyw\6WVJAEqG}?ݹ7h9x4gk_?(*IՑ9\;V3D4QA:Ǹ -(!9>V21EkFuh.Hp iB4%;V[@Ω_āx OtAƴԷt.c|jj:,offo5| MuPPa@Vo:X ヾ}mo/if F+kk3XrA8ϥ5^Y U$;?j>\:BosZN@We*F [v;<-yxQl?QR`JꗭTY)sэg30HIʅpRdA%8-=ȉe)mEhQO(IS?X┳c61^͎LcЎ )e,f؀ +PGig5/\ǥf}/ǩRek<5_t/B{gf:X8)R!YkKLL#1R Vc̔qB\`f6^ḛ@6rȓ$F2R:U`Jl;Q R5E'DV|h_p~&&_tR }B."RR}x5)+C%wۇ/JLW6yA$,#d|ݛ pA] w;]34W;CL2as!1YP5vn*[u[ЪTGD5;M5ptd N]j_ ;p.Vfg "M鵳V!DsέV$!$\A( IRL^TU9v$sMLjIs%WHͅs\['km8ˁS}n$ ;v^l}D-I俧z"4pF:BxI/Պ;C(2ZF#6 ̔F׎Ó3d1j9Dk Q#=ъ H0 hש̃b 嵔C T(_@lP2bݻ"StduaI+Q$ZS(La;hL@i[wod݋YΐY{j]c.hH&d 3a30m^) b )x{VZJ~ L9;5E.d25HαdRnyCrDBIfwpaLsvG͂KD1Ja>vRpڛ=f-rx la.HIbY<͢DT)1NșgyW?d=}K3ID!wCT"e4.Gw |Gc%ҍoo:sD Yu*QyF|zx˖ .ŝhȏ l::@W d;Pkl wxMJ|SO;K͙X?ooZm ZœoY(Dm!|`f_nra7]p9Krm.G^oGxGWN};ʫ~1{(و0+ܕ>9b(++?ӆʸv%Jˌ3 s-YɅ0W60tuo3+#v5If(k;9;?Fth>9b¨h`P^Uj ]ʒS%qr6EyڼL˼r-e׼6^#䴤,1`JMn~k9 *ۋ :ưxGS8:#PDD#ta:.'.;>O/g7gq$ev4Yc;@-jAziuZOY7? Ĉsʽ cSd?h.nM4uN+YC0̅\wsXKj 5a=kv@0K9C%ԂmltvtxC#=糛(({mņo9f-Is೛O=LSw63ů1H'UwN]hE6rS.R~W0,\d>"Azم Z;z歗CATǯ^ھjr@R* ȬbUUx~W5WKo^ A $ /V|z_e|n;B k4R"(:n|&R"F{LYE*F`pʀކ]&UH02@߂JGk!hsQ @t/O byiʒ˜`.yRcYD¦h18 A-8hXlt"EGQTt0?V9_nkY ICrwaF&ielym%hfշd!N 7w~ Y,&W7I @v .H=QrVÆC$?0ɇL>ea.>̪7D!,Zj , l, h`RqF#V uڕS}]>~۴r+Clc29;?^%J{xW(Pw|]&Uڦ.#³(ϸ88ŔCR籆҈i`Z[6H6Jgbg3YYh_e=K+%+uz^랆B\c>@`y {RCtO3@DOL(ֽlɐ!=E1 to߯9. ,+.ô{ǵ!@xq"J,<$ 9)̥gh_?0(:NM@Ѯ#5WhIڏҴý,f[.ֆ16|;-fqJZ<^~qw}0Y;-",ՄjcE8 8G vw#:x,DeTdJ\ƒ@UA"-)X"KhQ =\gx~GMX'nQ A;ĆYɌB aBpiJF11yZn_Xo+pLAe01i̗W>zbh68"xÁ`& ܣX[Rn@>xbV&8Rc`pl&|c)%7sZJQzo94w֠Rb9Դ zJw |)6G#FntUsu'!=AE$D }ڲM49}/J#K FGL3&FOLw6B^8!ʽ$w)o",@{aQңZhDH!Ui-ԑ](AF@iЏrdex2aZdy۪<)LUq ; !XSBUboNo"݈Kh\r.k/A. \zPk)xH{PGY{b"U=mO'Ăxis &鋲iN4V0BRPEa . MЋ#Rs!f1,} Rō6:}xZJuiZʟooJLy=[,ǫҸeea7uXoWΟR+ͨ\tvc 7e|j4Owog3O'6 2Ya.maprH)6Ŋ n!iVEFMo-,myV,2f .'g̯95+o~>0?UtQ^CFiBSb]U<1Q|pL/Ŏ61j*SMU^Cݏb`mJ ENNX#]d),.%Bi 3G'*OLxc1S'$Y։bV {aaeޭ| c*]XW8j4DwV*:]Vae[˜(;4T 1۟Jҋ׈;漒r'qWC5֞u*g֑U DC5,s^?8Vl|ǩ{(qCݯdjB!Dv{tAÌLI><4܋͡X7;5]<6yP,8jB}B.׉s-^]OEp󰬽V\K!50˃('!;j?]T<{]؊6=:dwhunv|;Y=d4c˭R_j1{}up2u[l`Ä`s ]u-mJK]h^<#NS#;9w;2pjlYޕqdB)R=XJ8P]űDjHʉg0jjiMV/a!"Q,bv 8':zq~eYkO|e6pp_%<Z >F5&I`tU(^ Tk@ۦV* ͘J;aX]P*s\{dCۛZ;*Z7O}nEh MVBҹx@oo stq G͘FV"Y2 tܴ JV3C`ZND;͞K-*͆g QܺD^Pa$5$ RRwAwsM/m}7L 8pƟ,]/Z<-rt|")h(K9㥲:;Qus2dkI(PwؒՒ3gk'4)CZxܴ?HIhܝ$*qC@&MX:s`;ilٿQK R&u˷6J]Zΐ$N)M̵DR#HQ bVK-=AS_]튨lt %UǒܸaclAW_ 4$&r& /Mhdnzο\xnlތv)}6(yx$ H>{~|!I f5h5Dh{HvnY)Ó@5ҜI=EdDz] N}=*,5ү[B k*t?v|䞭Q m+րms'{^^_~to/r=ȫl+zӑ ;th;kk흛 CІm a /B^x7f]Qd~ OP3䯺B#ۗAo?_Ŭ-xBMON*҃`N3`+ْ$[ VyO8Ғ؈lSly+𿺿7 %%j9JC!fBEpygG]m@ %q)Ԇ֚!ua+ macSfQ/C/PөEl~k@hI\[Y|syet/i\LTjiV9127^`ΙM fM\qBR|Gw`e+x>t7N_r:8 `<~1C zf{iԇs}!=ǥLUc:<ΔWhvmOTTre8YJo~yw÷oFf 8Q׏ ~Ek0X ! y R I$P a]XZ(2sz $t0қ,S`HqSN+ b'js)V[ .@V12#tH9J XJ $+0kʂ1*!;׎;EB88ByΥѡjZNQ0^9OUA@p_#[<`A!PX(&JJ9GC[:ל DQX$P@H*P(;W#foQԜ//3O8 J[FX -ʋ `0ذ$,h @M7=|PF&PxHŅ)qdF) 쮛j5=Xy>Z:Z+DVxjF? G32c4zɍN:1n_g|~y+LV7bgM=7륇4p|lV>N&JM¯B_*JOLђSh%47G7YG:[s"d%ƑX}ok0].huh/WÅwl>q v1a%098hL )xnI?N!Z {!Sp Q~:q=k~ضW(;r?Dr ?4vnɻNƣja\NKhGQ cl$5]2jHAWa-]wE:"$'U5Hn9tf@ur"cQsZl!)G-o9VB-5$9~@Nط5`1O?>a`FjckyMj60_6_Zίh6_ŠBNǦbq\K>c=87nG;/W;6ʔȚ!qN1!X]/pkAvS{}(kkFC\EtJҝ{7)w»bPGuRw/^VOliۻWkz&4dU4I#&Q-^}rD j-؈2wKА=Wˆ|NwP-:M[y(SwKҫݚА=Wb ֮{7-.xT N36nc4wpkz&4dU4I#wwKŠ>c_6ʴ yNޭ sM)Ivw]Fiԣi,2#ʔ#Խ4#t sM)vgwSh'[*uT'xLIwUr5[*S|o#\w'[*uT'xEL9̻ݚА=W$r{{7v-ZDب߭ǾrPwin hȞhN<7GמA è:@7j j[K_#T ZTĭaQksoel0";FM}xQKغ˸Ǟ-ֹg+l5j 顝haBsOi@QS-s/HiC;g* o`@e0yF1&:%Wc0\c5&- 2טsQK\ 1ƜkMZr5f.5\cn$%ë1 DQ1s ë1 BD1sj̢Z1E,"Ɯk̍Zzx5fp1s(^9@6טsIKd^Ԙjx5f;џk{Qcz^>k̊(k̹ܤ%(j̊Sk̹ܨ%pAWcV*c5-A 1p*טsIKDƬƜk̍ZShx5f-5\cn1=kMZ>f(ҹƜk̍Z%bp5fx>+;ט_YSTV?yԖo o~yw÷iY|(b4 ,s s0EDY'ȝTQbчpa;^?{돶^1 m Q&ѨHX! jhAVIk2|Sa'3" f]PzJjQ4 {F*c;!ЫCXZ*,(Ӗ I ΅?(k.xBkq,`TD1>hDU!, ._o,Gw-Tje`*BB!gN:p16x=XR-Sd% X;:B!`1b ;6aAcS8sڇRTBa-1+4%HA  s,qC0d gy;@HjBHLJk%*L(Ғ [`eţՔss2Zʫ8433rżhG7f\ G-MRbDqwږٗs3t0/c"!:zf5϶?Fi>޽͑(F1ۿ㛿+,[(~BP2Ax=AxRŘގGaX 'BoN̯j6[) 9"v_ b<ì{[xT&B!ŏ}T9Լ g5CiYA@ҠNV;R &b-uLjY,PJ F}hA8u" ? Џi`~?̍&|;ӏ |@_\t2vagQ>7pY?N!Z6|[؎*+qfN돿 `*&;Qt 1~2SunP45$L Ѐʫ.9-Dc4|Z,ƞ8@)L?M#]g럟rO_5pԗ\߿>&M 袌~4Z]^?Hl|sx<};ơY89e~XM@#2&KEs(d*gbQ4_ݼzkG+Ӑ$;4P !]L f┲ <"U\&.3Oz%h>A2.D$g[iܸz_Pʔx刅rY̧]#Q~G+x:'u{UE30ڪgʚ8_AeP݇"൴;b=:EJ @߬H6PAgEUefefeU1f& #(`C)x:?'SJk 9o(k޿}LFnjL7П`=NG`rQ 5_Uun[IKl6' Z9V̿)yV t4UKmuJ-}m6xmd$yhj#jn!tT 5FE!R_Oq ȗ@8`H2* aP}dR) QL7 fME̍NG,8V1R%p;HQiJg5Zc+MoWA|_ taq%%i!]o':,*'yQ^1܊/I[ڡéKJ7JVfK=w8X#YsrkPJ*  s9|ܴnC,Shl cl.,U}󻅃8\6\}XLSziAyF#D₽, &pL=vILʷ?[Y%[^MP7n{GIgjV2r>Puj N%N!K\uRq6TKm_2z4[E=s/.|lTwmkA6%lzˣwxO<9)fZn< 9Ǜ7>zR=ϞA =[♜1!< I_m<}T`Bs_]``?V͑j_?R%fW 4E7I2I2I2Iռ GÄV"c4F`$")̤ς,o ߈DD~NPrKǰ7;Nr1Ķ'2l\֓R*9v.$r&@8#1 -3:,)Goz #m7#˝~s(K‡[fѯXһ[q-zGؙ{\ jkF`Ob`ETkTX*;0<$`¬0bDYR#~7~̷M%&WcEtL;OjYA"9eT2XpLHR׵c7LYn'FfwH?\}<ݰ":].m^m,5׏F+C-z;/rZk_s7m) ߐKaV[N(.HfsY ̞>kͺ?&'n>}jIQ/o9LҭrgΌݢ@|%z&L=R}W>KE3O!~SR4Hc@\҂a M)˨w⬖3\=CJ7suNı$}mr # М5͚U7{nf72:k=NaJ0 I%LiĒte#ANٽ-'jכr5ZKFiU|ުfxfؖ)LQVFFAsL S+ ӣ&H$ZBL~hD3Enp='cLkiCZ-Hz Bݼ:\ )<]܈;SHhpbTZ]2 J䘤K)eQv{L5FD1I)7>eO%\\9GNvV&j4^>Z!%=IҤjgv-Hg`0}m| r>_^3˭_&?J^5fg\ C,"!p0>4\?p"ǯn>VD2p+y ϯ!@1 +(Pk3v5fV19eczpBNrz[s8mHA*IO 𾝧bLXHY"*[|z-oH؟b7(q[c3Syx+B!e3 ڽOU.o [Vrtpaw۱ M7ڏbqۻ6De0#q\0m؈E*NZl FYA&g,d: co1&?f"N=SPMfsg]@ [y=F$-j/mU0vv Vl-O[Rp\{H)n$robOv~ t`JmX\_-GĦݍ1 @FH]#9;/yf~9zYV=WeuK[;T}93'B=]N'( % / AxNz]\bێd_YHrr$-p{&L0܀!'`B2I i~g6W=:wFa!;~a$PMq-m a3g.v>͈Tpww>`Jp>Z036@>!g3vhWZt6N!jOi&,?d&a1=c-SL!{@v,:ct!W6]#,:v/ݪPQiFt*ܷҟO3Z9ԗ#ݻIrӓ'MOnzr$eNI/#J/VEJ)`Z轱852:&B2xop9Rˉ]r#NUMh8yrS,!mPp` p HE>VלYj.F-/!eT7_ܥ^LICj yz?Rs?Ny`.쎭:Dsrz3lܡa;TiPBh(B .j@d4x"'jb *\lC W3 GoC ijsBjM n6gɽ*RGͬ|RwѰ,7GA㮭aȭ_ʹn?CЊĐF7BInLhPioMxDʫ M`HW}6R`%cO,X^I$g,U>c8g0p#0<`&UPϦ>`tI F'` #uLo TpqQ,B[i":*L=JRvVm'?Z5Dy#_5& $-$LB2.$Uf`K8ƈyǐF 0j1`ye% \b`p[ sjDƉF =qp^Ab)rk8CAtvHdRNpxL@CqX?ʶnLM`P>*$T>NrN'[jp 7@Ƞ b ez2DXWQK&[q% E|5A92(*빀G ^p62Xs")Cl/}M$I9!ouf V-KrيnDoH 0WBa,@f逨bdeƃ΀ 9+#}f?}$֛W{⮈+;$nW v>dOuE0=cJgV^wLɃ.S)(PM<{9옂i  \cwF_J%ĿHJ+S1„Rye x* 9˱bi߿Hbjx4%V!=+ ̓3P.aX(kb9q(LTsVzAխ=$z.H۹ r bg0Y7Xh*gQqL}'x>&)8P1oDineT0~K6hK=A^Ĝ`-q3Hrb}E[-e$Hvm7-1Wj@kl&8 0qAb=䄂*Yx"ْyIA#V ^ |JCxݎ1L}t= e`E\D|3>b.Dq1ptB`}O7,MH>~'x m4Mw77~xh1qB5W_=|fm"5:l;'iً,}dL2-I,=âUFq 4^m|sm<(ģ^hD1RS#ߺ۟윌S8R!Q#* J;V6G[Eև@`0tvnCv\sY9 I;O&'?95wO*k\'$A\DsmL!֘Cu5#xB? R1M;qLR߫O7s?bFS;~º[I+k9]I骋㟵cY Hjx(X!ͱi/-5T KϹ&GbOb,5+ޛZbޖ7}o?Y7-,+9[]x쇻k0$f` Qb(dd$ҁ\w0K)O. >Bb,ק#>LթuQxhPHIq))j4kf<,-mg%3_Ff|̋6|9A!)/9.Xt GdF#ʉBHq=?D9G{-Fut>Xem׌y~[ Y ǞNdW KO@=kx u.e@ሸs$F }9\cIu5z=^,i q JU۰(uhSS˯[&@lD0Gj& #yP!5QKl0*cX[{1T3- XV3.Ei+-Hbj74׎+A9#A1h0[Fg %"I O*a*,hT+x~c\-|&54% ^I #. #(a7C^, qEj ||w1$ +?z5O<ǽ/N>Pۃ_'#DLj8ΧeGX1a9q "$on`]EJ*Qbs81JfWo\`kLc\'tόUnF,@eωiM$ʼnALVD"2TN(ĉ7bm42'.5"K^äpbft|Q EOΉN[G`$VT!4N(Ø&&d]u=Pqf*eט0c |ۯδ9bidzi(\gLz;waLS=nMPo%, fkSGb ޥZK`!ۗC#*Ι~Xq\# Nbbed2XȘ5^8p"^Ô?>l_$/+I? N"ٞg|4=ԸIc298M!%'E>OU) -SYQPҢ+D &XE/kRkJ+.es{$=UDIIF6EnXGްZ \PLj+ӂt,x4 #?9D"@JSth'hԒ7vwK(ڻ =޿_@q@fTST#u'l/y{"g]@MsԖ,!&/($bκ|'!T#@'K~@Q^Z3tA5k܋žV5bQGŻ>.]|g{mgw_jM[v7E5wY wCHbn?PT y]N*?_͢NlԭŮ uk4Q 2 8Bwy gTu!/MYZKV Fh6UBB}8w- 3ڇqk -aHwUZi0 ҝԦ iՃ!Mj#Pcː2?!ҐsҐfHd=kHդwof#us!nQXKԤHל0ffCe`moִlתtdAv"AA5%Sm^F}K:<Ӽ8fU"l;{OY<]Ujk{U%Éӏ;*4s;NK)N`?^-}*S*Pˊ|bK.7#lBbx;Ě!_jCo! H +qяoGo# v9!p-Ni~BL2n/2n/sbk@cG5,ESC74Iw 9Xt)X+޹w'32vE kg䭷L.,,jdQ3B T`Zrϝ5 Ci<++"Hо.-A' m,91A 9.Ak̨tV4;O!9QsEҩ nkNJH5&SD ҢB.b2"ΐ\F!nv%mS6$'xI<PʜiGrg +%vx1嚋 Idwr9LcJ"%ޮ+} a/Rd^qIa`BrsIE8*mI^̺bLC'%KØ2 1s#!9B3lros ȁv0rʙ$'mHW,آak;YϾhS)F)CbGhݢ&˺Li q]J6$oN:Щb!::W9)FE B"qv "-vʔi[&?oPlA Łɑ 'ƴh\'W2.d=4}T?xT˔dZ ޗa@Ph FG,=6Dt2v2}w{ˏ?ݮ(C\ov7r4jJst ha8H%6F M݇qٳ  <9`F269 eR ml}r[bZ)8lu02DHRH+.`3c n|gDfʀPȀB G3c y/p.vXY){(ЍזQjd"ZNDCzƌB"ՙ[Uaf$әy E"@1hQ@ԺLԊhaڨԌy*L;xm<u+<# + %)Wx|L'(T^SAɢݤ42.+Y;Ǡ}* ` gBF)jIB%#(&! D.@Gf8V HDd =XO]#0ڔv;DB2S$S1X.IDԩє*X : kpu%gDdp,(EvD3F)r0PgT(᧒c)3Hrx$r:(P!٦3Pg^dNqx+@,dӃ ,n X*!fI|M=bXWRm8nf r*bE];6M)XTcr)N6Ҩ8QBXE߉kWm@a}Z-li/.aZlHv /ny 97?Fg+P66W*3TMռfzwY5Kž~Z~5!pt'3;ٰӳӭ{yԧ*LҤ'C]=/qՉױ#6ƹVk{İυԢw\ӛYg wIrw UeHVh%p&T[xLzRJCj&FƠV<6{F" r[QRnYbdX|[~(͢,0J9YƓIDy ‘ 0:&:DfBc9z\bRqm_(V,Yb{ogWhj>ί 6Y?m~Uݭlc>KaH=\ŗnUys8:Dqy锐·5D`I=`7.ָl%v4Ԣ\ uCNy^QLS^U ^-eẓG37y9 @BWYO9Ts!kS:(Ix.Sׅ:e wV UrЈyO#U1`\00 l1n˜lH.|} ˎn۟rNy4O0_o{@"bo8ן^ .?E}lzmߏ?\,,d*w lιZ-d%y> K> %B\!·-z(9M$>ltXuDӾ=>kFN_m _y|<ב~?ūOfNfɇwi|ʎpjˆ*ڮZjEmxr[D3au"|?Y}?c[?Oe#_of ;?|ʐB[_: רd~0c9Pue'M?BBy):媃'C]%c; NwuMw7wN-I9/(; % nj(yƝXHRJ` &"&:jQmmQ x/]~C;cwu5TBTDZHRs%ELlc)YI $%[Ev(~+_a5JLHhD; '{x9JbIǙQ40p` `DOY ̖?u*B<ӑ6Zr:pacZ1ɍżC.HzKDqٗP$k׋2Ch .zx[Kbz܎H "ս(?=8 7o ]-)~ibUVֿ^8VH6"@ oԂioJHwhk{Π.bq/:+ ui/OKvoܭ::CpWxBh8ԇhA_gUpGoWҭ=U(m3ernhq{?t!D߱e}#rt׎X3"ހ|z|Ώ_Y73u bx GiHlx4IѾjP\Fce`%"q 3*ΈR=g`)-39͈XKYB@enl{f[$SnnMqaI 5#%NŨ'd\ +JMdWyR$ : N 5ujiG[ahKMM4> YEWal&3^b0F C]6QӃr.W^'H0% f.Xs [Q*\J[515j7 R~d&05+ձ?6PԬȲ>A;O'Y7c{R 9y]+y?W+oޝř?[𑑟XSDz LJk,/Lss#gnFIJTU,* gDԧ/:@$L@  5S)@Dr{40g) Rnؔ]AۯIVtwσBVgn`h`]x3'ӥ05?]Lrc2tфheК@mw (O3uAߋC0(it^:)Ruu+&ݠ@< ʞg@]ʷ碝= LKSUrN{:@&DZnΒվn@bc7-=(qAX/fh=snxQJ}*/İaDeeJx+ "kI^mY}((B!$H^xRB 9+5zڕK2>R .Ьڲy4\6x$f}rS QVq#x%XUD0[xOԠqv 6ρDX̗9e2D!5VhzkAclDnm.Yt2dp mr"LJX+ g&z_/x$^N%[Ea#Lbzk|v^G+fZ*j"1]~lp*|D?J)JkW\ee6mi0iv0鐉M$ÝN&oe= wޕqdB F}&&13"ZEQIᆵ(RjY#xtW{޵Cnc瓖> S,!CAؼp4= q be]+ktg6k6IfY@7|?A <c&Mk%78gsX~Ð*5#.Bp00u:@Y}E DqjrO+%7'>tJyu6+Zا ؄Y$u424\э8N\!pƙlpӴ*B ؟6Dt8IWܽA!A!ђh3-6R-{av!Ӑ:oy+%_ӝд$sstBpT8YvnR]mv:hg_k{mRϯaTta 2:F(6BjcnF:(3uW{\#+<ӜD`Ufol1U{orzpOTQ R1%1%Q?qN&{5vH!)B x,8rAF$ct*.1VQ+ /%Ҫkk`. k1"GhFje LrkFmT"vYǷ /#u_=^#[i%vTq`J\dgiݬrTםG`{:=w3ꯞ,z\떜ii ]JЕ4da/;"Bu"sjnBv]פTk)O޵۵ǹKffKE]ZSuv!3Jb{y}JR[oSY7tFRB:.A`I=lW>hMp]6԰5БUCߨUgzw]w-JΚz4w{wvs &{:z:kSG {wvx&j]#"aa1iTVŻ+7;nXC͸oj`iP?j?yWCi%⤧d4H %ޕ1V_{+</P w]ߖRd(u ߄.E|o`@-̒l*5X*Wl"VEpwcUpbU*O*V \2NV맲$Q0ZnZ+p_e;ݟ^wLJJs]+3ROSr[z55FKs]z%B"νOWJ*򈽗OΏLk ^)=cdk.ѷ#o V~ϮT4/V8{wԅ֕PہZI7_LEެ} A$6G ¹첽,taMNRAx/8%<Y*~ |3}S\sn T$"UM]r&ޤOި-F+{s k7n!Z\Se[o6D?#0ך;AZiԬ |X_IŻb#ę5bD )A@Ds!/ %^;rշNa;˖RTQ,&bǜ1yώM '&GV%oA$~/?Ir,U$ژ&y 4۷}$M2׆ݹLɌj[oMԥP_k !T]u 5Ь@K%I˩~}e7Q0̾bAlCVD /|UFܙMuh%u?LqIwX|(4nC9xvG.hDU Zj;: pH.T{"ϟ P.-A*u @,2m'[ORl|*lJSO۹O42=Dq"!\ ̴ NHA^|D1tacc21 B(o*8DQCD8*6rrIDmw{ E̤4r"qțT֙4(EA:gAS |&mb.!#!햯T47؊nϮSXF ,,pA5" e^f2˂Ej:I"*,MoA2籵q Ѡ] YCtL+@R6)l%/LP.Q.qxjH`Z8g ^qJL ^0mjzd86ߺ=\rQPEA)sQoV+Ytڧۓk.#!%H^AM|_3"hOU&GуABI8-O,G\a*%&W3"fX-pH` |7 (} a䌱o7~F,f&%S~J%e(x(&SY_PCzMKrXpd ş, g9)3+ZhݤT WkF.jPQ;Y*" Bڠ0X,+nOAېx1㟰4ø[.@@M$$OROn6\.d[/ԾOG[*PXX! /-]@5ym#R+Q * )l=#|ǭ7Af_oh^._@h6ǧ{EyC;yq0M/ɤV:7_47 3wgc;:Lkc5ѕ6PlJ//_+wL^MiZ|?ZOuGUbk-Z`ztl؂bO/at:WD} or;{j/ Y2_\&_޾}s[O#_^b ^??nr4v>anq_lfb>{J_S#;~Nِ NlQX9]P>Hqj3xzb&Zo<6>ogZ8BCpd.-,M)e1CÇ )Q !UWWS?)پH5K=?Q,?W~3> !Gg]1:?.Q/G磡x,r"bV,Iy0]{{apGMӧ'מ߾{s?ˋዾGrTN(A_y?Q#o>r>~u~>%\IW?NMb+oGSvjxgxy~}QގbK/? W7wx_(%ں|z";/~ngPZrMKhmהy34f uc?b.S ] ?W*}DfYH39nkr/Ǎ# ε:%yܻ{u/9MPdq̓lwjr8ݡ; wPjwjͬ-.o)'GPc7ytÜ XG(6qfiYmjU3-@PȠrqu(%\TtJ+uC_ }MC_W:;m.Fܕ␨.E^$ T,T9[ :hANJ6Ld:IP.Vf4UʨX][G(Tί~L~0;;wO)<\rE+հ֔1x9C '-:؂V( yEk< J8'{9X֨hteY|.Dk~0dNxgHT0@GQ%p)j"9:H" 0wܺڻdmjKvЃШv,DT^UCR+r? 3Q .KU3 L8Ýh-@xJe|yHaŠ\JrF`Re ܕF %ȘKv=^KJ8kd"2ƃM] ]mj\\Z qW@7JbE.7,p"-*JApКD,xOe4J]zLo\:wukwɎF5.\T}{^Jه.Wf4fHm-Zj*D;$rA8Uir`"R %($7gXEyJ.'&T I0tmj"nh#2b]E~V-g!\{4bpC Rzr(o,"H+uIQk)p[4ESHY`[F[P=9c[<.'?~90o?? (.7C-?.'/.ٍxYtr`'h^]_VrGeg?ۋ% \ |p#6{s(5d*8$If`^w6É@s8OnTuB47L׿ln]Q3eqaDbF4OIͺ5 6-j,Ꝉ}5k^D7qDto] u\vm%z8(QIU0@B)g,b:NDCDIIxN,!gp<ԘL[mCU s}rEJyGɤJzo޼{T=Ć+>ԨAˁ% wN'{#U6`eЌ2AI(*I"Zx^hAr0%cK;/*xV%>t.".g(MSmv,ڋEzg/6 P,|;R@٦m&gKj  $NJGX@EŵR!TXzD$xd yj2U<$[M€}QeJ X zk &h)#:ȔJ*O\HH6T4W*(# }% {Ke -ѷVP oJ w-5u_r™LGypZLu9ձ- L\Q%*Tmk2L = @xf?|4{8KW-"8 jr>O[HBwa&2-pFi kFdh(kt <BaD(^96PϠ`hP2K-dhЅj/`cKUp:I\Nsu4Ln$s =9I)#Mr Ϣ5ؤ}|I: Y]:]jlFm4j$}uT5Rmupt{|(ҥWi5~hp *DB/Ph4R+~-C8@H JA7aL$/* &`"%% <LŒF@n U!O)3G \(M҂f Q@![yzѭ~>A$-}U%}Ӈy~ J`O R윁+~T <]c`côFl.9{K,tY\^i\kQDksmOT`Ugt'Vn͙SZ$>"7Y_5KHn4kiӬޓfm%ӡpetu.&0O% qߢ󛧛gE?~?-N9b99{}_3:I .#RMdrqWjf=+ C;MEo'?]t}'m,H%hA^J4sAjQK-*]S3Ԣn^čDku Tp%6oq*ƴ2]-EUߠ(&_C-*3I-*4yB F-RKvF1֢259"̠~vYvO(+Pݛ<ȭ\Ih%+)sGڕZq[K7n$F?ڤݙx(rG%z>Gh u1QƦU61JJryIrH zH\B[b;s47t1ÊQnaPJnGrJ*nCA_#@󻶡0D軞yE,.DyP-hH|8 <]c #4' ))@2lucG\39h5O&ݛKJirY %5+,&rG(grk:$;0y' j|q%$Ap :˭T3oyIdmH'̙IV)ZDG ({ll'FaŇI}V4:)e|EQH]!rF]*Ef  ac54L0enaRZQEZ5"0SB5p9o}S`n1-n4߇ܜ> RP&[e(q'CA 6g[-kAjaA%6:0 sYGtJ ! Fp"/4 hbNRrUq|T1z)!q3 h{X9g#US"67ѐ`m%1ۿ4u9IF~ ;@gRD1&Ò"3i@61nK\yk2 &n1))2t}18K|y\+=Pz˳{s/>_9Gf>M#w1.N*HQj2#^~#ЖV# Ji\٫r%:Pi-`scj(DϸzBVp#;ے-RV1CvA(ߜT\Ȃr͠@ύxhݢN@"XtMa+Y36 UG8P)sɨH+-`;MsV _Ƌe΢d !2C[ڧ ;H>{(q$TB+ĠM5gv Ndx2YA'X\ܮ@2 ' GsTi#}RP$BLs:e,Rhč 9Y3xTN{@cXut fTN; nU7BJ94p!ҥnqC@iWÆP̎rm64 g&VL UJ&aCJ?ř i'% X[&($B5KEYjͪ,2(dŏK)wk k뚤wFnBstmv8䐎#҅F ˖JF{e MhWoCF>QJRj'#;7׼mӡ3T Ds#cr)kh3~4PuBZ-kٶG5Q'l<!\lkkt`+P 8nƵa-H>vI +=f8B t6>CϦ @5+H*J:yٺ6(4NoCmIVɥүyA dX:`t=ê4O*׾bzcjx-{͗N:TϽu8VG\˓%2xLفdc͇_q*M9~*H1ӘLX)LʙjltRƽ?Kav^nB'yJ|,=%bJv?n$M9lZ*:J iZLIpرm׏mRu"r=8 ú9  dD EAhlxxbS/-[{{cKv \N2c8 /P]5kCtZ3jER6~]}?>Ym,.ǘb术lzsɆkTiKf H T:nk/)mkCx2 [$mKo)9T#ʘ}iT`9|ͱ26@Ls`P5Uqc^5]~x!dIDן*;i'A7-3f-:s9rfcr8#sK~_ y֫:%JӓQopގ'yJ)*9䐢䢁8wf%JsĤKBсϔQ4/|^/kc0BUR zG{S NUoѓ@}fyqoZw5ⅰF[į! Z:אֹ5\^Z9/S^Y z1kˠ\gbХRXFDH*) oQ+: }FۑvuoRj8g[iljcnRmQㄙ,܀ma܊&m`op;> +dW 1)hD,%6\05g$IAyi pZmY7%F.0Qڡ\Xz;:*acixF^^6 Yi Bc{-QBLt,ΠNY"pv~i MNlww|:!X ]M>V"$p(7J"g3^^R#wRXY&SuVx%2^Pus En7iGbm2H\)2!!zkPܾԌ@m$xF{YOJjm|NYSiQjhaqfcI L|z4H5Sfm 4rӆ Ay3IpMͤ@ŵBtt쩹r\c[fάG#REc-l61/ 'dccM.SfegZ7okDKgl<:tǶwD"FǧOΥ%15V֠T;4']s%}GpVD,(WiJilRr$GvǿiAwDZR(CXَk_{mlY?E[ aK7#Mj`i7 J6-O6LnS(k!mhը،2pvE QUZWkF˙wKg(xӈ8]͵slgs┕u[F:M{wϫl.PܮlB1 D,$)%>RG S*Z!I&f% E3~ dNRKm\Pbk$Yw`PyŸp@ńQpU})\_u23Q  h sֳlhFI+/92++:zYH9r>; 2s{l2UN=6;*Zv?k"^NZdS[kЌ hz )`j5:4&J ,ը`|%hѹ՜KZ4] 7kSS92omx5װ<R4#=9EՏ_G1/ܣ}8|C){?=?991J<9bQl1sTt|h|/E zV KPr&?8wр޿1oFܿ@0U b\t#`H"\LgKOh89v h-_(O%W a]p :%q:YEfB 6q{w"mp/f0޿w5y#x@ʒJU #/pK+ wuhqG[ϣq%x.h3I~c't47£3m/&ffe̽Jz t6="UiӃq?LL?ziyقɖߵ|g=}l{^y|pϟw'|}cGO |u]7l~{4˧滧2_ z{!];iQ(mNwOǓt6vs9jGoFg~.GTN}opf->Ҥ6Ko/>YQ1W~gF|z~va0\ ޵m$e q671b" -[K~դF$Jjikl$DU]U]' evibf/eZ^`v}Ml5M4r|,^XO#{gcqf{e;ٝ _{#X諿`s\/'ck 5',UgcO>cxFx1e]קql3x{Y^Ʒ`I?vqt/{ Lkyz=zq}Ɠ{Gk?υdr}OsNrԿE'ϳ7E1({x_͋4C^:^" j\_mjի]6SRUƷ)YbWzYڛg^Zq_P[Ԇ{ك,?-t ܻY]V\V6?ͰalPDQHSjLEjV+r+[A6os붆MM$|%0;)`KUSc% b~|svu;}(&+QC|L +׏ ATQ톨vׇշb\ M m6^)5Bm x IYp[f}C  N-0gi+%ӶinT["ı>un͠ڜke^cƎ냛a×@Q c7ckcB;> 'GDG,bR`dhw(1IHطbL(wzc؆!4J%iLC8 N 1٪,%84GXSGh $:I&"MqbnSL ME`Bi (GQRM%)%^%Ҙ8,4,8Qa,L!`"y /|L2 EMm O c ) ߸o=Gg~ 懈"|+GPiL}a mdznCwe?2u2,j;fǫW/ni#/|]~E|}]>~ ; c]e?wί~|6#?%տLvOL/m{x~W_}E2I|"cƅWoA3 vGd29 [Ap,Va;nYՓջCh G#hZԭ7HPТ3P73.vVG1kqgB:z6&ӻۻ(7ݷ)~GTz!X:>j%66ii(CGKpL( CŌ%4.'(Zk-" FfŒ"f:BPVNơ12DV1!Q&䈱%)I"p!BE&JM0ZٌgVv77|KZ8֯pjj(ѓ&]_/lQL'2zIቲujֲz 7s9Q\LIy r Co v=()f%`q5#jhIiQGm^rF4'ԠNU~V1KG*Q vxy 2iV O x|(if <*yЄG&_'Ձ 0;8c85$X]XykWYXp{1vDT J(` ! 8ew2Fl'\}njJ Eخg|`h Vm%G%Hy:6lŝMNMgcIH )!vP^,PBޜ_ [;W4Pg/O=>|p6&?h+t5ɰ]<9>'+V:4]t3X2VR{`;PT,,9AlUXt|Ɲr)DgtE$XIS%_eD[PYi_ a"SYd6"Ia9-%?iϢ/`f@AD"RqRiG(! ShDaۤS* f].xuKκ7x\0d=GPqJ,{^,&f9ͷևr\jD|¶ﺤ#MIREGuҐ$`*EaN8Jt5MLӭѓL?mM\I.#xzlˬx~zSzޝL;! cBDB0Q{W#8֧˃a}a Qʇ"C%hs_OC{~lc-/O TF{7>nvj4KD}CPfho:cҷ-ۨα|)>8+rYbú:#ɿrq.o\M OnnJ,SM'v ^9mY9X¿1G HR"#30bcLsoyޣ)k ),R])Ӥi_COX&Lp"IU[U[dנZ3F56TJ-SgpUX9bŚɴjTR-S]|w[pEAb /U6]HRlrfǸS9cl``.pDL%RxơOS;4U CSO;_ ƈs`Nq 3 w ̉-G=aHUꞷVc̱~0Hb V'?! Sz߃j )? 3x'^j +8] ;=5as>ch1Y E Kâ)?#rNdgKaYn`J3>#(s٥38K֩grkqfG= kgs Եbs+ͩ ˯Յ$պ+jv١!$e(P?H˓@Fq4Ũ#ŤgdTjpS5}vVw\b0Pz3$&kQ1h֥q ^ GEđj2L*I] ބXQ%UR(J.֮j8}-N25s3,uTRoEK V"3:wIzWf0+-EhT|__Ei& "Ip*Xe\`wvM)UhBf `$AJ,H !* Cی%ؽV ˵Kw&UA9#9N(CBi4i2K Sx%GۚI}Uv~B8=|3\.U{f5Kjd]ܢE_zv4w//VE?}:7-T~0ESepwar#˼tZ4'geaॷvpsK&ƏO /79K̅GuBF}e8'rv+p %Ͳ@1Ѵj07<8{H5=8="9v."Y3표1zHH0!xU㩮y&ޜ*n]lIPxlG%5-Q'# vLZVπ\޽{*5tޜ~ M ib a$YB<$Q&= 7Q,l.9%69S$Չ=0G!ё1.P_U蝸 Ot驛P}U?>I:IKL/la7 7}aPovlf9;0\nf8p7CJ9'I,`<hc$;p#g|[&xSN[Izi9AX{!~wu/lK'[ֽQSaLu"bԌccTvƧӋũwC4SЏǝ'?<X|:u:|w`"`'7˾c23A?'=K{ }?D>v~ qS)ο }3<{tQ[ϯmN&CɨǗfG? !(?׿_5I)hmQx"bJҶ(d$wP2 ǭ:X'-~S6X3B.DZkVwZVw3m/'&-8Mmv,^ pǧuȤH`Fŵ=+0.Um d$|;ىT1XUNTgs
ӓ]PH@tB1™@B¼`IX"Id0 DQ5^K!ܟՁ@aԜ<=,Ȗ_ ֒E>A{uZq,q FkR,H-2`l1*L8N!2fS/iYjCKHr `ͨ!^ 2JՆlE̳3 F-Ls/b mG}Xõ(sQ3&6mb+Ե0\ րóy2WAȢg &y#+3aO,1xD@T V`uC&^b۱SKcʂ9_Z҃)'tOGP'IG. ܽ |܁GD)( ,Piɜ""T"$QBOQܠ/1w27RII"#G`eH-ۃl bA<1X2҆q_X3.YFir*+Ş  +F#@A|Y3ȃ/6-&VjDFL<,eBl4> U҂O`>А#GI3%g$` h=83g!Y" ͘xi)<2ltrfA bX5"=!EKB JP(3 Cܴ =z n:pb Gl.)`%#B`qhs\KڐY;#m\o “sSͅGJAkSgBSz<飸|ƻ4)bmoysLuDs*Ѫ9n1ZrZIZ!ɹ _C<5 i;Vv\o裼a+fq d5'/! ,;4SNusj/O ònͨsIeWIzBY"-R`i(f;щmT?_Q-oVv+[4_Fj5FƏ?>D諐𙷃!jy4.7k:5҂]Y]EE%ޕdo,!.RוMQ X5&`[@y@Gun8ڒ~ WոųBc盟^}61w70:7Woّټ`Ƙya8\co:7=+P~ۙ=8Jjǒ/ Mޞlt;纓6f;Y7Όqӧ_ W#Mޕs^ Udq.aѵNC2smK7􇽭* \Skd~yj2%DhMZ>zu-IGZF9.(JݞIi{1CХ/ ɂ3Nm*;}UPհ3ӕ(5۾ zhT+@ڨ6ϑ+V1:} [7j Tx7>V bAX C|6St-UnozsN.nÇ& 5?RƓttheCӄ6_ >> v^2b/Y/C7[j?lOgS(֞K.Ϥus`/kK8m?np3(]!üZ :4 Li{Az5j [vC,'s:~>Pן5JdͣE҄ \u%8Ċt*M @[ǰOhּktNVnTa!#Rr&iq6/Pފcox2OR֜C|) ^g2WÚ1!ZmUc:{.fod+'@>/"}*n$FD\ot^CW|1-{X_/w+\LZ[VN9 2L*cUxyX-O[ _pB"BYRʺ\]GϨ&vg5G׷.l`ciV/*jY+VO?$Ma<슗R^+맨*u}i9-tS7?_~~h< o Wq1AxO\qf8+h%<|syo% Q'3qr9Gp Q0D2CDt[23@VM|Az%.k©,98x&&'k՜x|uǓqJ.0.5~L棺ݟ_LnL"xn\ʭq}w5OhT6xL{ .k޴40grz$Z@"hIMz+IQxY0uS", ~ͥ9הKYg_*trm֣Y2ïGe0e" SS\?Z̚1Qdf^fS22S"{T MQ.UpP|wU1Ewaa<] ZSÒ)vbNu+uʾ(w<7J`Sazᥥmtx;ɍHsy+Q'<2SR7b)g&j xKI2wҹatnX/+VN E@aKCRṱb":%"^y9 Tyo Z,N;P$oU 7HpJgz3K;Hv Ċb{DUWHI֌}{4({ EGL^jjN>x&ueS?nN&FzGK y" 紣OҾ9Q^Ty&7Z0[lY^?. xB i* B"0EOUCxHN̐;ŋ"{ %Ko%{֔914\FQE֌ߌUM:8zM7?|k~)ܵP/51E[}._m%"\`םJVCBCF6QzfvQ%kHi27- %O"cq?tGƅ ӯd]tVsK^(w~qu|tTlnS;Z|QN5I=RyW^K]z_կk4VQjM#kU`3 hlb?{ȍ{RXdHOIvG<0qc{d;~EɏԲئZ&;[bWX/v ::Ib o냚Nx[bxn EPn!W*i(r~ib KMP)pOPMhV~6v5:$KY3\a~U(s0e<[֞m1OCcaxaBy>{?; A7Jzz<0!XT1`<:8vʆ? K6?xxϒ qgųC~'*J>vc} Q QsؿGckA68^< ~XvHh#3@{:);p(h@7v?8\T >  Fqҽj-mU[__Cx:7.(oG? %kOlO_, tH. ~gYwֿϧ> RgMM9, :Q(ORAm}>b5>nj8ɵprklO]}@x巟n?|LaiL rϝ|7鿝,'isn2i5sq6*nx:PlE!ӲtU\N_OwtFudBL>@ ҠuM,S1vZiN8nO8j0k9̻0"9eU}*D?cH;K{iu"d6m[-͉bMEӑL;0^B:2BXg :VؤY,@*9̉(,iN+Gd5m;QlIzvz[z.c6 ִ F~y3ϜzfV5PV5Ά%>DFDTʬR}e3{8iT ]R[AD)h6Teޗf}6Cd$J`א͌#i˒X&0 ˬ>]\uH O̳!>O4Z0CX_Í<] h.-"IF~O q("ok|x~)ߝ2'i&`r'O?~u^흪Ӣw\>JY퓋wT`Uӓ_I,^*­.Gsq k -WTj ؗY}[=CkFCl0&jOOϟĚ5,E+7)b-eqKE8!Yߥ(b.% y RW> )`.e Į( }7zc;PF93FMmp${fMeVf9Hi}iU&n[v9/BNxvU"EE|8&,GYh؛^:cc  ;/_mef:ꄌ҈TKk>6F!ɈO~~ݒId\p@Yw4ʚ6Mymr;qy~{0.9@SfS+fFIn3> v;K*T2&b@d`ƈkDQb)#Gެ5FHn":c N(rb4A%#9bm Y)!9%#dIuM4UwLo14l !73T1bƬ."*"8 w2; zA![76c78:B3v UH;>PKl_.gJo0I;A9.RƁ5Ys;rB݅v(Ӝ]ծs%Wl-J2,V#oF9Zl1YdLʂ`Y\V`֡awCtb`;#UBQЂ5[r&:aK{-8F}qqL؈bG^, E/;vBJDddl)u*Z3˘AJ=pi|tRc9zz1eI+d];uƲ~`URy2GQ!-:]HFʴ&qcײ V9> 8l%&cɦ(Bu%q8`Cv;L6#kڢ2+瀭h,#$~*u#Mv/\ςd):a7pٲok{Ů*JMIy(mhJ@".0cT @+jeA1lcMʯ;˶dsIE'q##Ÿ;4hvHYa>²sN*W3b9|2:IV ہ59gZ(N+d9AGsjCiYmȐ/\ymL8W79c2vhڰ{oc*i`oc;F~n8޴3ZMs Q,F!+ђ%qr!R#2C5JY+4FȬ%ِ@iBvК^,d%QݹKC&أQJmZ,ʉgԹ Vsj(9E-͐d-p#̈΁UlWׄ(Ѿ/3|+pCFi$;ۧcbΛgfU ?X nӗ''O<4gm>ޮY])uAWC=nn_^p_~~{zy+e ~u{J7xy17P5m@:!f"ph6ڪ]%Я3i߰Nzg";(i34>&ˀﻇ<]kH3{qh Ys9FKzy/?wkІg=HQQŚ&G #vjk#eپ7.x}ѽɕRT0;) ʸV{?o僕{q?9xs_KLfE\`zW2)f`ف-GC%IE{ܰNСM=a?e:EggWI@7ȳڻiwsj1!X <=19oK=g/CeS&99rR;V2G0*|J*-;Xum/+XhAvѺ$-;+IQ@ˍ 9Vl~/كFYvl/A/ Vt'ﯙ'=kF"]ax/hư|!?hirMrL<"(ݨ AAl(9 t# Ae^`>, =e![^oe92 4zfi\ոZPnh#n[Mٺh4 u^(hזaCq]s;\CMI#g>2a Eځ5l!gTLuH#سDg?-N]R+mrK.(dDvD,C)䔌G~`냚F>|(Q쿹O#^ʡcq>Rq5_%>6Mld4_Aަ]TԬz3rg$vk'bg^.VQV{K2v4}o GK\{c6BqmFIkyh|5w\8QL25{fS7i#V>w<̓҉W+v?Xwd1n'h~={aΧ7ޠM;EOIc@Mg#!0I<މn? :+/t8^CQsN9@5}WSضeҧ4aT4];-PbC:veodSSM\;^)l̷&<&ֻ7۴0h*c3Wh!EԑUZ0VvWOB F6mȁG:\ U-$N1:J5e;lBr zQыh I T+t(?[6N R4pOIJ:q&kY4q?(ٱڬ7Ȟ%-ftmAuy!DuSɥ9ZH9/C6@2rD͝<,\g)W~2G* p~͗]ܷHeѱ@~8_òx1;ۿ $>}'Ww1|j߾^nRA3,GWLVxګ)zXtb,=z;߽(fe' ғ*ާ+bGrݤPSWnԈN3h> Oݚ_VBB"r"8hFͣ4i7o%*F9i|i6sS %]DG3Q#͘=Mjڨy<"WJ)`8r}`4{#Z赻rR฻w]8&?M*fe9i^w06 N-ɻZ߱`Bka,Fch6eM.vF J톉r;`UI=LvJ)&OM/KN+i%v+XI.%9 '~[Ia^vN / =$'{^6lrƉ-{$ٞ }elj{<ůŪbn69&?mY,2N248&Ž.!@X$v sFCpv"_֚A$w79L$Z.W%W# 1to".˄pcE}{<LGщ$̃ OzB[ԥdx6JƹLݨD:gNl@+ֿل~yC|+CZP%͖]heaB QQkׯ.9*jcFY 3F b9c$>_ ͞1.AL$JȮj+78XiiP4) VVe uiI鹤U"9fK3a>gxQkWV|PZ za9N0ߔօeO%v4 O[;'{`K<;K}iSLb[pt&  9ڈڭ Jj؎3(iCE ?WP;x7D)yL ޘjZ^;o VQ2&\aPJq ْ 07|Qq,>( ƌSx\/Fsߞ1h&.a)+-Ծ`T`E56jm5N qS1ouŝ=c\*$UL{sIH5KΠ&ޅ!ޘ>\8jZ7>zJ->1s=:OVDnv>څv0X9V SRK?P\Lr >/M *h!hzoO옔Ӫ2ȓRH$)qqJ#ӥsRѽ=SJ:|HWrQrxL͌ZtlNtlNӱL3ꜯs\,)D=Y$YU QOaK6KŪ[-ө%t'4!DP֋Ġxl쓢^>`3ϸnԅIJzO@+E%,`:RqQaA tFhN(IS?]ϐ@w`}R+jq5&+ J}8(ơL̇+YVHhg) -/Ép'˰HbLLBFQ0.-֜Sl@0u2;~Tp>i`N;HEieiؒ#4߱0Ɉ*cVTc) JPU.ZY_ { qB G%b K@TKT)YiU ˏ6) R!M%.S$,]HY16֥}49+*Lql9cxÜ \Qpe%euiAoօ' c0$.5 KBNp3 m>s8lvL`4#ƅ~]LT υ1:\k(mMGs]z:LܶwD0: *Ojbj6?V@/ 4 rN* "wKg &k1]n:/^)@2AUwOV?BBL.˄2e].xI O1.%Avgc+_+-2`BwZ9\8;*]=w&;S4f^NX t,W/=Ԣ r3*cJ-؊vX0P={zz ow{?𮿸k,xP> 92=c ֥N5ŀRDXw+֭yf⸫qyA] jflVA!Ԗt.!9kMY4*TKiX߸ yAs-sߝ{VZ'P{h}(*T. AEk+l+xyjc>Fu;7UXgg#gHNg6#9 eZ>df/556Y3xO]v2,eUk7xAʶ0 {WI"25V,x>(D\s EϜ4T(h(|7c!4K5zwɇ{4DƈhTLN3ZpgHDCL M@\I'+4k"1J+ш !_1XS+E.eZhy/"h>SKnc*#g#0T; 0 @,u@w̋#'T2.jL8xDLٖ!Yb2408i0Z BXRge+r|̠h^U)'bFI(_N'1hҙ#qqcLː9@kC5l2S1x"u9C>63?'1(3F\cGݹJMF]լ;n_Dsttv7(5zzH%˧;W螣!Pyx{r]H!99*$JGmBήU: NlemfO~^@nEd"KԞN's~twyyika]nuY*)g< nU'(2u@! A|Tf&Sf]1hc&,EgqE xkg PjOcY']JdW\[wҩׄ0) :(g;8ʰS3oV}XuMbSTYK3X˗w6j]˗qN7J!Wb*Z괺u Uxpfid2O^i W@,Ga/55۠L0in`7Z!4-l)z Wj [46Wr'6Iļlul*NȌ;/n-W|$UQBATՔ7+HhYGoz/h;lvi%e{/v5_5tP h^/QbY"Fr7Kֲ_3-U:]Ǽ ޑ_U,SmӇ,,ǴoD%`KIG4-ym SNI}[|LWVKFˤo'kG `룿Oܽ5ת*fTKT$qɋ-I9HX^H<ٴxcf@_q&BM|d =crq,3#fK $WI:6ٻ7r,+BtrÌ=3*^jk&M)bQ)PL9v=argu "m_Ff6ZWīk. JR8)TSba, ;uq[TE Zخ$ȪZaOSQڹp/vv;1DAӴ~IxbZ<yy7?yQTj=оz=A)Se8H^|!'L vϝF϶҇H Fd8N6.zL:,36埞lzwHR|Z^E> g;cԼG{Xiܲо{S4~˛ps 퐈|JɞzP$F(q?kUHH"H=9 G=RWH0dɱ3D%FG-D3ؖd"%cFؐoKLWֶd샄CKί< S{/cҋJRwK;"q=zNɹ_8Tk8zwG:q?pGUâ&ps6k* ,qh:XWH9;=jguԙR;cye2rp# qMGJ"1M0~uC%xL(K\8*ɖ&с\a#U.ǔv ip(ص APFX3PEv }EGAHq10.# $$H9D!8Nqb9QWtw HwdU9`2R̊~%Dٮ7Ӫ Ykh/~4p/yLH =3qa$=N,jNE Վ̅s`f85 X1Vhiڻ 1LTq˸ųq%⸟yiGf9tY Ag@=^[ǔ:2q1f IvD^)gP;dcn# K^Rz`ls+c罍#hᨱ\4rpl 0@YA3$Dw=a@*2*N(Z{ic^.$s$\)zf(x‹o7YlȈQ9.ӓ?,vw:nީ ȇo֟_q,}wW;"|lXF}_swSTR%9,JG q&z$y4y#ulxX!XTj\AJn@aeW݃H٢Igt:^\µ WT:9g oTՔ W%XNE_/# 2Bg>PLaMoo™C9l]4i+QXw GQªy=& }1B?%ὉF0a!<`|(G]S"iH{vZu'VzYka^ަŝj%NH-kOnWFL,J62WZ֟'I{3pCkzd 8vGP+uSo6O o6~Ɋ>wkˆ쟧.ҵzp W%:7t(EiwjD1Q̥nCTǡd@2OoнƠg 3yC2 +&l&9 Y6m^^G^&_'#'?ˬc7?ٛ۩yz㕥fb waq{(_u$qKĕ??[4);e t6ͯc/?)xwnRD-lP:޻tN>.|(قŗY|[(9Y|2cpZNP(bwbTFEgNZ)SJ ]5òM#1ݎ1<wa݊o^=i+A` 4Ée`%*֓{08PiRX3^Ijm;Ql|ZٮI~nAD c&<QTPy pm2W)T@bIQT "p_ų`8d!(¾ )^* "  ).&-~go$ԁd;%"L*l;4]'SYLS\:Bjalܕs]SJJk{?H PX{|ՙ„Y9Bn ۷+_2wsw}rʨ=#r`_C_y2*=r֟bYhNjhrH^$@gYzhDR9Qko~w\[f=\wSPHQxhLa^YBP?_1*[ٌ. ULn߿慭 ӳ[.thFk+]YNє}rSx6>.m"R>Kyv't'0-XUlTrr6zW{6$pRY>slJ4G_)QaB{*Њ(Z85UrǦ<~2Fa쉰'Vkݪq1&d嶾6PpsC~R w9SC_FఃJ͇9)Q@}oxH> AO{(DkkD{H3V\JX"C /@rWHDAA:!{1ai a;ESVHJ󻻟U~G"_]i6]L6KrZ-2dr#)@1Y>$^<0ֵ&aW}]F$ɝ>hD~aRD/\"փPY(Fv8hMT!#A7CQ77.!MᢛXŲ;LK?_ϷbNZU}!"^pNx4()Nd CiA,P'X HUV@֭jd6UGnRSB]|z~E\3F8B`+4yM~#Dz#Ьh[rգ$[F4 V|~݊"W~YV܄B9opPONjtڀQ%5G;n.[Wukv| ӡ8w0d|94Ar2o\mS@Ei**kpzV56<ߧy㗔]`ig|L6 ɃIFx՛U=}H lT=,ZUZX%g.e3TSX^n:{5G: L$zj`/֗2\ 1ߗY~Ron-MIT7[j꿀RP:G*d5h,"-l0ZaCD}v꿶/jˏء(N99(U9Q DHAP/ ,6*}v|HC%R1g/ynj4D^Ιb}y^fC\`/> ~Y/s1qƺv1ؚ=0jw;_1/.Q5g?ۗ}yV4 EZd+>ut4Aɀ xB+&}TX-7rJ7lm%FOW6Ɩt֥3O~1YD!3;ٜD/+^=9̵r%߳*c?}gQ7O G`C֦Bd' 8%8lWsEy{?|l>Dr$dwXKMU[:;kAv)-}۲#az )z&I3MoL~SɫZ(VG!xXG  AzB,f_h}Q /?r+~O.H*4TYg2x Fj.Uܬ(!L0t۽v:1v3%AmkFCέ4 F!e12\:$gLՊ,5s9H&zǎ!i%ڌgIiF)@*kҴOԈ\NB0FPVcqQ,쐨>a 9"Զ0\C+bNۥk Nw4΢B5$jĞSf)$6|Y,Oqf鹴nvA,zů⥾PV_Q<bmQ_̈́i㡚w;ռ?A^kP䇷^48QʅĖ=96OX <\˙c*δ? *zqm۝p4Sv W}Tթ 3 GW=^tc[D۩rfc`FaSZ:E .>13ciɘ V2Q>rwЩ5QNA5 M1^2Gm; $p} 9oӔ(;]:jt i&=_0+E.fZp90Q)*qN[@X4BwP|.wDӎRs;} '5L0@T0vg=GJzuYx3΃ i@23iF}CC-pJ>5O|U٠6u9.͗97*C6߽?KtߡP&l˙ԌOOMP,[^?G8n. L[t?T Um?}~i06֓s|[6P&?6T'gR}wS3zV jL!U,ݚ;Zݦ?6*tpFaBӊ bcL\;l-R.CtNQf uۊUpyC_!\ĝXkyk\~EʨTٓ8dw"B- i NPd)ٓS| (z:P*@X8Lh xYG<kA)FMÔ[O +cqWBC#9 fJj'* 0+hVMWEq!>]Miɫ9iGRxf?=NCGa?VZsn3!׿<Dϴ^SņmӮ׆Aн}̀35-1B, R9L*RʋHO2EcwDpI,`hŢ>TO3{ZVМ> r0#"rˍ"S=CehFd#,W1u5Z9$By~P55'I 9~Fǯ9.ԩ`?y盻o\JAc`K F'sFcrp(' 6i}Qph'T! 7tצN-u,p.?(g.G/>/msTy5_ݗ_0: f~J(*cv_oXd]qJV[y&XZ|NV+qTUK5_޼JWʛ j*,D@2ks>Mˈd/ !9MXjEJ;-W4y)n7 Aؾ`o2iPݚm#7rۣ!^[t3֌WMC89@ʙyrMپI?{>*顡l4GC)POKaY$s`џ}ҷ⨲w͵/w"ǏH! 9X 6.rAJ*Yl T} NSC]zرWA|bpЪ]wjyEj|,paS i 6])/i;HG8ne&<~rQXwvQXpG"J6΋=vM?O-rԻjӋkQҨ4:~I Cb{ksVRA4&žAjcZl|JkN3V=m/&kYL=޲$Ac lV[`"amIY*V4O0p ;}f x}ya a-1lꃰ[Կ #Y#DHYonmA{ ;iNPj Gi4cZ9яT=_5:bwH%MK7IV^J}pLt,oMF]so.%g$_`p`@\c R^`ufWZ&RΨ WZcʚW҇\CM\ )H3r}_XVoBY RׇR)mg zgõ/điF'IFjSi@r>1׋nmib52!@2' s\#/S1+N d–zvd_Ju:*Djީ uGiݹ2  /4xM&*lä)(HW",LJ)'*؞+ > x4zX k `:$M690$k<"Qg:ǠSI+{n#9p/0b,uQms;GEW"c1q#騌d !HQRH8ᓤsW⾼{j_˫\MdiTl3sƪV[i:7}C!gZ@l c el"e{5ן߯#k+Vn_c:ou/"%lJtӫ^ݟכT17eqwUv1OaBӏڄ6vnotbۻcv WuhzC tQfzB&dSV^L}ݤ ֈA41Ȼ/"XλvڻM`!pm)թLwwC[+5vUż[sGk76%$WϜa+BOQzCש(ӻr]xz׽Gse'T0/5i M_$pPn)(Z3YKROPQHc㣍k&9,c+d^2Bwt7 L pE0F @MP@yaJX/GiAlcǻOl3ɫpz\< z>5&f]-z26C J%P+g}2܁ /b}A=m\0R"_X(e~Ed0YZ\yJi4ѥR(Ξ$۷7o@hp^s*$H&HCU$LQƸvs&ʺwr+ Y%Dȍ" ,g,y#F 9arJ#1hI-Zs= (Ȫ9zS9sJ^)mtʜZKlY19WL꒼AWR%1͘JHCu 4*ask M#/rqPh=,cnW844U~NK{(B5볲dnf\K=63 Rofh-=afHXmf3LS6X̬;%#yYog CT2s]S}7taחW&Ϫvg?A0#K*[ ~˃3&bS7O?oMV%CZ0[XaT=\!<oQi1:(\8(5|x2 1}E|ࠆmex”ˍ Sl{6F䦳_e%Ջ`I3}G9u9*&V2}6|jVl M3#$#duҪ`=@LӉ*7E*F7l <7<D% Ѫwj-6Hm:;j l)sd sM)-׀¦`&͕IfIfBu&<:3* @;>X(t$ 9uphqa-j8]z ͎ڒXw,-)=,u5<֩zrGQAN1TGݓYݣ y"Z$ST~n0$v+ GtJhݎhw@jO4Wu!!/\DdڍQԝ, GtJhݎO%LO4Wu!!/\D)AWF;k~4JFi5yLhe%VI~~Єo:yJ!bFu}ElL^}\#.HAWLGӴ2tgFn$<}劎j\[wjrs[|,{OiRM.ő/|]oAnC =~B R`FP}ӛL(8$6R >;2iɓ3m;I,~P: 7UFGPf@WHI^_ y%FS{?kOuՑH ;K—x*ah%Oa Keϱ|>x=A2®F?c"5.pj\7MPζp1GpnE5m\$wޟ=Ӭ9~.i>+?)&xޥet>㋷?IύM#կ. ^XhDT'Eheb`35/a6_>P2|/|M>H,X)XSvj{wЂIM6XMQT 9樥(Hn`H7v6"N0C"FA  V$)#)1 Jb'Eu2`xsV"&rVya/P$eA)#Yǭ-tiV,m ,!ᷱ K|5ڒU4yVy$+#|4+뫐X) kq qXʤ &hΜ@.&ځz*˾k5xVsb5dAo.542zBO0eBV5`Ah`+h034Hc4;'Tzb&$Sen)9m!3aL`X,atraXi-kLG MJ`\!QDZIH'_cgXgA%h(SEp`sFj{cUΖ/g۽d9^r9s^潯KkZti>ƥ@u'0Bg 7}VpFW( ~(p.8y_Gq$ONmytpm9F>5#5-8ަ)^k 8 6^0ȱH@r t#,f*?r?F[T|ݢzLӗ?R_gw8F3R,ℝx\_k49CċgwbB'3gWj&Uӯ |ٽQ7q:^2oL6/\k^غhl0#;m'./IoK;%GǍ&KG^)۹ÎQGj-|׳`,%+ՎN&|rjud~ӂVAŮidqiU >\o%G40u(OWr>(-=!w`.2[ ab8V|eqYum;fO5?UЎ1NQC\Coa Q2`D'o=mfiHs^\шrVUm1gy.zB{<;|sv4A5>!; rQÜ5 iDh.nx& U`u@2:ϵD[dcbG "RD(i܈h'ףFE)B1|pKIj(j֥ҚhB3E!`\{ QB)o3mi&u>ЩuxJ%0{" sjd{ # V"BKxqPZbJXjfT`H-T Sל ptGkZ`ȡ zX!B(.9oC.`J7πSO.P-rϨ&5|Rh$ZGlS!*D S!:؜'L7դͫ K{,T} <`*V/sWAGɢ-)Y!/E4 HgاYWU,*@MWus .i˥m`n'qv vrGG_].NtbgP8ba;f J+p׮JֆHho:\Ǯ.挱]Nց5|oOwOſclZoa#_3NOR/fRff'(5dTۓ,qrG䣚|UB`L`ikD݉_kJͦbls\ Du.Z>L+ 1dun\ -?^ٚ;Y\PT39S]փl\S(+a<X'Y`rbHR]=Ǩy=˂n4]*Xz@eFF R`sTh&G(`1Bbd&pZ#jpn%]59lQA웣&{ ϳո1a]3I߉ɬg`bJy<͏̼q#cݒyu?|TȽţscilvfQQ<}>޽^SyP獅[2Q*$pSi'.-RQG]Ćs%R f# zg,ג88S\ lC+VsgG;ݕC LO7,vv u:(us(ǬX>4E|Xu08( v`dw(|ġ >Z򭘬]ctA5!sۅ[-hvYk%ő߇SM'Ǫ؆L!6yNDD} e UL$3Xh]$f`޺Y3ߛ>n>}V^^.E؁uz #SMwW̓e/볰tO=6ʍ,E)¢ʥ?ёp-)跎MFJ11hcnoڭ y"Z S jqHCV&$SAi]G> `XDQW\1#X\f_y|'RyIJ ?1T M$8 8HqN`ЙDӾΟBJ[Yf;veTG|If$Y+heLۥ,1ì0FWKX;15,BFU ;FXmTw<\Ì3GZpKbp|vD"Z[pT.ڎ;uN+ v`AA@ij m  jrÇ_{dz}Il`bgjfS*vȔQizhMy(fazR XR9Kz> 3@sU 7AS+7UnP6p hn:d|իڹnC?MO~,UDގC%㼮˨]ssW +x*Կ?Ē)P5_O@lx"Ƀ`LePʲ}. ˍ`VcHR0y- F@׮]o<?13u(|xQ!\Ld8M[/RS[0yaE r4j'?bHr!e0T [$Ǘ9WΕE`BqhbK[Q RΩ*`pcc]6I)Yic#N;`51hC^otkvͺy[ Zc) i̢p*~O,\(>}sђJ2e(ޟN%̉*O $r]:+}eV-.h:ћⓤ-UNug՚=Yx|Y8SXp.,?o8am.$eqi*KTl+Ia#G|ZDax3m=Nt#VжL3x[HE&5䂭V\7lScJgCݓ.f9 iOˋ*dj3lesbLMpPY:0dRhf ByXv-%`L0T3% Cn4IDWB/po`CP()>PZdTEUR(0rEC(A2 7i~ _MG9BedmBCȘYD㛛&H"d)}7V- 63wT8,*9evpb-&(Si)/)y(3*( :DR[X{+I4gc IH^M&_ı$Hy)jPGwָĻZ[h K*JzS0 Qop?Ze0i9wR`m׏|MJTjWx"hYVff"gL1guF9iHR`B"׹U jK QI4, `BqM "]RWlcwؕ#+a眉6N"Awim: F|/ .Q 4*L((_l}[ RFȮƷ7=uKQD"%0-I3eitqgg7G術>c#ZP|fPńm<*..(D~aLa;䀽>$52)0/ k, !έ2ƪWVvx_z kz_LL͝($Z@ YlvW(% <׏3)kr:%t3Y &Ù[g;ا#2rR9:wnnDi;0:^rϱrJi*0+ 5$("9.{+* WL:'e8oQ oӪ5{aa&W;ONҠ$ x Oٶ\/ 5/s3o ;aL4F=-?yj65?TBFF}h_fhĂİb N2%o(ύr9d(N Z  hՈ7;eLnIF[:5EZcS4w+ -`i Ԓjx;Ĩ$ FtKh[JOr,܄n}eSLHbt)dqj$Bi٣8<3rX|^-gƩ^GV`7I%8f ,xP^Q)@mb!qj@Yqj0G Y"?H?cNHdɕf"9ždVZQZ[º@!YsW%4;̲w=%ۿ;'aK!0I%!i\`А g״Nx5*ylqkA t5zkJoǞ ,z09ĒFt2hE'XA Xg}Pp"O>av! ߉At3:)ByaZ,NZT0,QnjkWq#2NVTA <Kj|p෫,K)&[8 $Xi 1ƨ` YʼCh-8 iƨAIxe) ],}(iȣc^"`T18/L)Ed 'D!qUH kD#Sڠh$lR,gfDj)e>T #-+$ 79H39YJRۥوfbq\2eaX30c,<{XvTJ8J hfL8-bI`R|˩R"Ϩ7SD5RĂŁPII$Wh%@e Ҵ\1)ps̓_ l:r@;U#:ҥuE<&SΈ;CQ 1Q^bⴑ[' 6;yt$n&潪սeIke×%7ɴVw6sb\yd- 7 b}W BF`0(be_;,p0HzÕD^R3exu{PR9J| 01NϨ$cYuqB/!|wo&nF4+d1/cZS{M(=h >\^ G?oE?[x2U<- =Owdkv϶xs{}6APS߬Dpi츂IoX5 W/_jA&S?߃q-X=4}OWwrg R5~b LJ߬wL$:,꜑%QNoO$St^ L?,S~.p؞y$>QRhjfhߧ$ENk 5?ëZ}30pyyssۉ S0uaҕؕV0TQ胼=z.b2,n޼QYJb$ch҇R[!x)4nLq {S~}|s =".PoOs=ǚ'4RR|aA䱺V77?Ns BafZ(LsԞKYd;\ ? {Z =O:}Q 19Zj( 1)HlVvcL6'gsak\l4%U{t V;.J'#&lY()>g1S-cy8p8uMCdZ:GOx?y`3sT wqjgjkoNZػ«UHkA^:jJ(/{WM dv!QPHeV!\iIWNpZG]1*$:]ֶDn={s (D.o>OK;~RlD-V_PÔJ S95LD݀3iͺY/粧Ǽ_C03{p,dݧJ,ّdLOnn--}:&X"UŪbow5<V<>0GYD5g}׾i kߴՇkt/h7hT3Dݏ5'(o}pm&n'd} ~Gi=o+ i):, vpڲ$! 5=.o _)#Ҭ"(;8ʎa~t_2@7=| |e|0y# wGBh؂᮫sZ\)~rmuVC6gn=)veqNCBֆ|li}w# Ccȧ]%ݪwC/IN"9/ţc~(拵ӲWw o&EHӎ|u#$w\`jfO) U,j^pxo8q@cFF_Z?ZmŘ1GRAs i.%6ܪ ˕Hy\B 83 ɹ"iD4Ղ$8(\K/\WL)Ɂ&~_2%vJ9.]vx V=(֏Yr Պi ~(*^9_JQ}f@/=?jQ:mnhw]]?8^O}]zw= !͎l/Te\!s*V=|Z\hgmgok<[}Ur=j\MxgstΓn%&g~w;ij:9M) +eS ö+2<%ATb,|v;G_}q +h8~!T3gj߫ϴ UibS*`h^F`M.A2Gmnl\> ;?~FQL&,Ixnc4~ڣQ!A_6ut31Ssr2 s_fOo_"=oX  Zo&䥻s~s||;dE+W@BaWwo-3W v^%1 ]Q*:VTE>Ძ{`uy^PU赴 J) /rєEIsEq D, JsUok ji)iz^|l}6;S5נ {Ů^tcƂb lV4npu&~AV]g1dޜyjg)3=@g"\4-,PΣX %W:MRւIKy}+#F9@% W([[W ioU]b|ÎuX܈0.W%c@uLRuRw:XxmF'0w`.ׯu(Z*v>cƔX+PKR&'"'D)Ѝ0Kc,&@L*Ri}:Iлk$Y{)D7?}>@ ž[&_3qJtO6t,YD*Ey|*^fr/L峇!.2MJ_Mgd!itM $.:*s؄1RA"6gI~gU'ӗL{6 -7{|Xe6y'y*EBT4b9"Rc%,%w) fev3& msKVZarE Bx|^cЁ9CX4T2V/6[EJrfn33$| \(3 so@hąKm #+J(Oڻ,fr#B_e+(v X=^TH^ȭt!V"N8NJ;>~gQ5my/IF:G SV Tx )y3%Ik3x<$(iѲs1_-{+X[?V{KG<C8V͉|ѧ+D >Q(Mġd]=! y鮆$\rCjKރKUƬA,nXq;/s9j'5uѐP93Xlj~ǨjWU^K+s`FHCy?ܦB(Hﴗs8a,E]nٜ歗j@)}s.Y!qoΫI|!*9孧9{ 8p7ǡWudJk!BWknKD*9ңD2ԣv|Rrn7z]` @inun/c5SA##H)87F6Tb=-Gte |ר1Cg!\`cȣ9.ZUp}Dp\R@Uk@u7Z "j>g%VCT4)L^VI9A-ṨKNk%]|t1-uWz]ź#=.я^l0DCJJ.|7nST0#Q\kOfՀ Er7ʞW}*˷=-u1iA8ɶlriq7}}S)z{ߔ{{GaRHkE)ʤ:AqDE<Y4<gikQѿߏ>f4OoG |̊u x*t-g=圣]Vxq7Γ%DP?X5a K)T\)*o7-(&*U*Zyy=>I~ǝ::#)d2pϩNJѸɖYrpxc,v]l]-byL!Ezb9P է Ҹ?&D0sΧbXSN-Nnyր[~K!g0{`q.NI9xf-gdDc,tJHqSszHrmT|-.Ms@+0(ikP=#=Rx4ډLHWu2Eb 8"QHNS츻nW Em E_!hmEMSp'[=ѣY]##$n8E$!:K 1۪RJ`xd*k\Mk=K1P%cxq.T#񓮖2)5*tq]r<mS!ԉ*mTp]vw^w6%DdmeJG$J>)dALc_bOl!wyƼl=Q<,9=RnNek5r/6wZ.״YVtVpռk2&]nɨ ӯ=" N;Z'Υ2CxΎ"߬Ai3pQ~jKH(Y .kM$+"r]@ 0= a{<(ҷ<2A@i΢CH(+/=`t=Mk Jv" |Y)1:v҂'RI'y5pawKed' _l\` ^ZI2΀Sҝo<^NFUC-|&EL7λ_w;oH+0qr gT]VZ\C0+թk;eblQLEw%r:`Hp)MBn7sMG : !h:߼H=rp׆r@CA7؇`3Lb4{Z^d֡]vYQφЕ[mAANVw o1ura |VT^Z:.*jiXUN^5 ,Mmu,$lX1u0]MI*dYq+mF/v3AA&~yA%d}í-P+1dljuX<݂t4~<]/5FU4P= X? ${^I܆,/mr[˲z' A\b-8փO11ÕCT2OM֞s32Mz>Fy=x7e vk:;ǹ/f ?_Ƙ|>(_'_pi0LkO~KT7ʗdNW48.l[[rZ$>|^Dwec^.so^z}0͚| |$Teio] C525Z1#<"Rljbx}iHs%v20w2r,CWt҄L'qS1 E7̳+#o2JQAE2T*("/5,u܉| z nO =}Ky{sfi&dbb "MXF2y@4cyX4T7[RBr $*~6Ղ@7(6Zd$u\@xN{-Q(2U"X$K$ R-3%#ȃg|A@1eА6 d%)RkJCôVx!4wkSR*CW#6aokă+KZԵSIȽN~A P dL|mH7]Ǽ ZL7gM d<ӱx JcPzG(h{;GBd5z. N* ̡oHN*o>G޽]~U0fCHk` ""A8yZ(2sb:a1Ays$owFf6~gx h˅W)ۃjáNIAQ2>k c\ST:EݞOlk]LMz0tN`dS6I3$" [R2-<.&5Np$ /_z9\*y[|M˵ΟG[ ކPxJo%<dž)"e 'E*5 +FuM90qÿԂ\pYzYqo}xA-u\JEpwKUjԢpڔY%W]}>LsHpk.܏ ]aTI!Pޛ<@[B<2C;nW׃o]:ȸ|-1P`f58*bah-;:1o@SӞ1ouvi;gj޳0)3&R.?M,̖pvK(J2?p;cϣ|;DtK#txrS<ڣ$d<O<ưl „mbDsR-Y:p1BpH'8SdSkeWe}-iRyPcjL KĘIΐJ]+N/C8G11)"Gr=UZ.B_뿺\ܰ*dU5,O8~h1GcXʹ5V0Y `DbR:بL{ԁMruŮ[C4 Qpka5gZ)NRnZU,xe=Sq C*@KP >[J_<IMcNИ{9[K3Y.֙9@aYCJuGs%S1"V%e@hYYz@+Gm 3EX$FVn3s%%=]f@.Q]S5?N&>[ {?L~s2)ӚK0Vi,\pXR0Â>skA5n7&v'l7̵w L3)!X(4ٲungOpa2yU}ѭpךi҆# htj:Fk֢ vx W|5cUR"nKN+WHN*|~ԄP$0:fv.9fK֡Iհk̍8(k(_TUɶծ)?!"*pŦ%y'%c$*P+.TU!\Nd;E%Lf?Ky}c52N3ϒI8fֈ& ,k\$FBmeqAGG3TQaadVK-R/iZ'<6BvfRʽ0J0Z-;.vEBj8ԒD'&Z̓kbd1,t݈u=RH/E%Ҭiϧ#QxkmO>=b!W0zJ6ߦµBM"{>"wwU{GRO9-\x048794eI2x8FON|~yv)ۀ:(?|fûqN[5 ~Ś"Ū+ i*DQkK7ybM~-܂΢B(ZFecK [lpF/c* |\G/N>돹S0Ȧ9yՓ|'uTJtN.O?"Zc&uhh.1&3K:IZe4Jn"&ܧB$j^y<}>Y~a.9nnjT+-F@b p?,){`tMF)y]a!bk<H8Gb.LK2N'#p6#$01aιDsΧ%2),P̑P}Q4>ŝ9.A.jb8Rc^'/ '痢8q8!P:qR7N/H/# w2Q+:,(>vV9 ɼ)m1u:Z/~R)faqx@'-qw}"]p0"/$v] ⱔ͊$6K>G/oX D6qL!?H *n W<\k'x})ךtrlqBWӳ $pmYdo81⊣:&wGw:05,1_)ڸ{m*igZ҇U"%%s!yFq{^rqZEss-HpG]2G!W=eIp{H/]c&ܑV<5w$?j/|u'Drw\^k!8Y4:{U{=ܠƪGM6`zlf7c ~t1nx*<*X{tmaEg+4, c|%%sY% "70HK#"2s(jI a`q:E . \KwMz9ʲќ(zi>aG30ΌD'sH+35fD֧'J8=|o+PIss|-Ŏ"ub?Ahեˮ^=QKW_=!Żk+z*rt{a2j+_cK|5w`S\`vqDkrt~a2΢q3jq]i*k-T6PRN-qnm?}vM0.,&wyVn\Fˋ33?ͥz}ANo.0 ʛPg{eښ۶_衧1m/ɃOv:ǩ=Nx@tȒ"Nr-wrwҾnO/_vLڷw|?q翓 4j/yN[3ST|nKk8q{ia?j[nhpL.M0U:ۤ==^1HڷnE{l[?M\C£dxcu`R=Rk\>ܙI-xh|?>2 ܑjL+ Up rI['/pv7+(2#>rgQ&*T>M[{Hu Ti9@ԍov=là!0ynZ?=z9oA_(I{It7ό0D{a1_K2t8>tz&_9N`gyZկM\9Цlwާ$O:y{uou|bI\v} _+TI:Oeem6Cˌ o>7!'sS.*ʥ |FnN0AЗG"4hG6D V٤õULЇa|q?gfSfz/4Lbnm~ቃ2{C߇_zľaT hymsStX@䒾WD1g$~n jPo7ǃr6{8oLAs3ZRvx\ Ysbvݐ SIy}kZÖD걊*44ɝ^ Zᓔ94㎳b*ԻK(5E|ӭ̿zOAseȖ:+#zsrkKѬ*X=!"VS xdn`+KLdάo9:YnЎSנA3gh2ذ0`RE,B$u蘚J$PENCH UPͅИDa!Qtu,"33Cm1&dQ6jU-s\BQmBU g.(' bҫChV] ?#X9bdG\ TJmEH̵T{)E]`$ah 42Bsđy^yh(Ru ZόQY!hj&:M9}cW51~{=֧{ɏ:K^՜o|" kL!Lۄe&ٶZJsv&3൩\DfI!+ɛ%ś%4a/^fIeᚶ991DzT)'Mv4$mBrh̰M@/ %ݞyXh;DŽmcB!h%|>taPWKWžnYߏ\,uwx8 [r: =p%E -RVN,x?wK@9ebv6ǃ!D CߍY r#,D/+U8l}Kcx²r'z#v߷x9wB{gRzk@C !Hl7eV~k>:T+ Qk݄E;tIDpBW2kހ~5*5o,x0ufq#D&E\vA9T8X 3mE-ɵI>ymx(24 Cu! yd R1F!%JD( #ToUcFnP(R ֔xD?䧓$L&O?:seH9e%D. mpm{`XЎN7+WG1<aҟVtM) $i; W \>RnJz>G,K!k647KΈe(&R}ΌHІV GO3aTV|eL#05LPaʸQh.͵srJe#JHIDbSMF]aL`+Ha1Px QH1JD!wŎ!C8ea K? o-bfM24 Bc@ A(.+E^-p+SS%յ1wDW AwÊ;]&3Y09Ɇ3Y/GU~HnmܝԺl][-Uj^!&"RsOԎўZŽz n*VMsb2ghDA V"T1Mc3wN|Jq7fcxh /- aHo- \9UyrjkARh +ej^(r?d B%פ$u*+J0B];{Xxz3t}g38g[8#X icpi[;M<2gv 4Ymԙ,D-4ʩb+pO7a@1$^V|.h+k5D%) u70uF6 !2>Ǐ~5os t!8wF㍉6`1P'/z(^'R·O<:]A~73˳7n0IҾz? ڡhwR(ׁAi@qo% ߳?mFh}$*A`/igbEvVPٙ7pqfYS 1cN|S8F/:yk݂P.%%>+.-E8`TW ~&0wC/>rj$ 2=RMe61GgNnr6__C[Е M"Y<ީOrz;}Y$$ӤƹqZ#a#)3w[w|acW2H́vܶ6onNR'YNNzCi=Uv=R|5`Q8[4EDH BhǑӄG$xXX3r2H3ez^3Bۆ d0q/(VDBlΘ,7:[RDMB6I齦ڏ׶0^ArYd]D||:9TnZ2Lυn.dbߏ-Ɋ|w "4M_wѺ)he{ʻ9u%JגpdB\>/` w]zdx\]pmJ1k޶MPЌ9>p;ʏg'ǿj[Qޙ$,?/5[MkDT:`ԄrD HK^KAIL'Gfh d13!IbJ(aYa,ƈ6V2,RHj&1gLX:=x{E67s精g֑`6„N &D6 F9tEmDi5ex˱%-H7ㄥqTRcù,R IYr0+pL,%IՔ-G-OT]wMLz/:+RT( b.]؀@xc;`E,q[GLֈ8L(Bma$CG i._9DL?\zrz|wBdiJ`Td[! ER&{#IQOq&A)0:IҪ^B.>Uu""@84.+mxlA2W!D HPuvſ5aF05 iw#31Ukaf7n͚]<`~nPdiǜQ' H LKRt\AQzt|r\tG] 2C:ab]ӻelགྷH0tqd4 "( !8ZXPE$eN I¡!RD\sEc֘(1Ɩ-#Ao~iȝb]yI @Nث2q QroVWaF<$On@\R(ޠpq%I m"(CAA^FcsjrV9L_ފHRnFjO:m&?4COCņLjsɈ嗡%)2ƍcjT{ @y^4"Fc0 b iK2 Ka PJdZk+K3k7z"QCs4k,&l ߻G!֎JZB>Fk֪|9O'-A'XϴV5u?KW#4?{WHd/6]%oREsiXϕWD jbc@14c+_>*+_vVbWKнw Wpe3 u*\}opU7,$0R2Иxę3Q#Q:0 xGz%خu b~pVRZbj{#o 4=v>'mT܎gC[R N5 ƃ!< BXO5|` _6)Þd.in,S֘h Hqب?J&%t mzax~UG{OlW V{]h@Z՗@ܖVxBR:qYN e2X -OحsA#|7|]\;d5f"gUT*H+W8 t$S6?%tu+GLXeb;! ShK`Y롔;JBb\(Ngg\8I-0Ɗc !8_E$Ů D (Q9V̠ӑ!)MG>r#r@iз퇨EgZ۷Eϩ].ljwFl;؊ 8:ҍqIK`K#ε [-sQ1"ħ!|f?m,K(P#el}نʡa.0{rsb`* n[X({lFQ ضgrL8!R cR"<pbuzueLK"/IL1pbzFEt=v4H'gTDe] F8^$nD:ϏiTEn9 MGWd[s6)ISd3 ita+ Du9Fs@ZFI \iTMx5N|9uJA[}_xiO[I</ }O ٝ\j x)#'E'Ғ!,\Ƽ6؛ @,k$9Èfd;o| r\l溜-yx3ǧOW(l`ew6 sF:H$$MYJ 1(,؉m{`mlV@qRq+|WrnN&siT(rN" p&D9:)uTX6 )k8Dõ8 %sry`V2f=H"m+AY5]2:cJpqB;7>`kЅr4OP_;:?@zsȼ]%[ĨfMN89aEb]%1 - v. C6=N[y6f{۩=ТI,.BM1ʡDغJvh;\J~|9]SXleFeRFo[`hoiagh`N954 ̈r=jihvrobgt$;`}q󝂢dL(l;h~#7?vF4=Wawݺ_r?+^/9_uUj+ley S`>V'&})c%xt=V}ϯ42;"GgZRpCS\>80QM4'jLic_DdcG{_no&uw=a\w?3p@/xj=Y/gqs^Kﮟ>oPŋ= @\4z-xI!fHG'p|C_FTNFόN]zZl;Hx۰ad wlm.sxLnof7a:Y}2[܏ҡ=:JMgp#̰NC֞=JT"X nP)>($uGYW9k*g ^z`])>GdFD+SD("W(J( G%D9bR|ׂ:)w-&}ʤ(kbu!1@O]e%8v[8 h7_.l~O˦&ܹv*W4@E2 RX T)id*AdS' JԡI#rE+$׻6, ūCվEmۥ)1rtի-wr \2FWwm3fBl1=]զ@ů1xYkض"OT>]l0,x5~}ίGI/#\*4:3\f,=HVhF4|&Rmers[ dcvXV+0ns:Onfe `q#9yjQR). Sz=㠔~NURfemI">J6hMVep6m1v|0~r^dPICM#飈]z\ eFULK}")$AR69lj~0so#GҞ>De(NviF?tOpُr"t` \jD 4A.aP=>0`K?[Ƶ)3@z 4SYI&\d W4'-O7dew,-lO[jТSkU=/ˌ26JD*WT!yiz#c?]d ۧǻ'F٣Qrɨ$8lv8ÁbGCGj!hpY6Ho) rҨ@1ؿ4@k FVk2dh}-\3=& ҍhQӍҍCnߐYر Hl-JE -r M֛,l6G!Y.v99m"#9i/ؾ[?,]f7!phW_Z{9,89|takgzqo>r#(NGΧ6Ӯ<8pAsCJ}M }퇯W]qdYjg%s*H 31`lZh$7`.ؒ=Y 7峚,UCAu;ju*h |BzE(o]$"Bằ 5EL%aUٵk'N? 8w|7^=igm^.gf$yCZi6s/he_ъnF[3UVSbY?3:cGgKMBԫPk4:PbG'p} `TI&ƶÄZGjmɴAuLj-P x?AyK1v޻먫r#;u~6( vkI F]= X/WR<ˬwR( (* o\8I- [[k¿Cv]A-]xOhoI/W~" Ѳj=|aD jBɣMB&P8pas` &8f.yVւDз/2^?;- jtW+ՃڸCɭz;Wk<^ [sJ.i!ym僷*Q]R)'y*I: "R l.r 3;NX5"W0-)bk}юi@ڥ]灎r;FNy?]A_˷~i()RLY ГmI>6.T@'#sr#I w6\$5v#|2*'5 HΧybKS19[B 5$vi01ň[uY"oAz)W_T Wp E 配 S;4e=w)0=xyl/ْNvk?Gw |> bVPh$`IV)+.D 9V%B9GеvWN^ۏrFh_އ_ZhWwX򎺨n(m ~u1IlO!d<]]A/EZ<ї.dT_B)K)qRxDЂB(/oJ){UVg@8wC>piń!4@Rє,,Z KS=ɿ>ξǶ7adkAg/gX )Oؔ,.JeẔC;4bc{-lTLHwuY㢯}%4Gw.Pvso6UB4Pi\tA^\ak= kUT0tqxd=5{Y;ň,0ƽCGBO $ ʀSHm)ht-䔔Q3FbYzggrjc" %3"#ŵgv uJ9ČtZr;eꙇ*/>NHm5('ZPMd>ίob[.` rJHVsJu{2rADLJ͝VT 1d)'J#J{q# Z6\MwctJ$8 )V| 5N aQ,9L 'aԚn񣃮}760ĖfZco-ŖqP޵6r#"eqvd/! l^60H6FI"[ՒZ6[n2mQů."UҚ!7 ffh,p"jIDJ39'\jB+KfA4аbj\3䙶ރ,4IFAcYy}1<#(L.-3b,љ0B܈) lO Ӿ=AKF~C f<Ui[.|+UŶ /|xLӻ+)-gb>; P?>N0>XE[B0C70 ]9~jX)x(JK~K rݨV$ZpJj O1ҺUD5-80V7ն^CGW .vA%?7GDJD5Cl3rkJPPޘh&5D97RPPMf!b?ه_ ~Gvg߽sTJ3e+^&czd*:h8]A-Anwf\ "iي`BVl Cgba,#9AZv{2@ٛk.#VcjXf%ƄgN>"X qbbSDJ,6 j=n0:i)*" 5 _IXPl`Q&${ t Ӷ[2ށh:"U>.h.z!$h2(ۆAA^lm ys341?%qs(|\7f#!`FapiaէޭϦ$EwS.EOU^^-ūAj>-bo?uR_0@f\\apgP9ROB;z k=P26Ŵ EX;KC0ҕo3vڀ8<'^kPr;}VT6դGr+h[RFl'WgĤ6$zƤ(uδyJiXrUc;gcg%#>{yx/>8B`N sz)il pIGtgUbk|1.w# N@el;6Cx3*4(Ԛ,H!.&5ܔ#}& r.k<tXJsn4Dil9)-%%^:@a}F>|l:TsFK!&[g\( #:4mqn!A)7H2G҅FAbH';E|,,L6B ypgv, iAN Wb~_ f /fadד޼:I B w%z&!;'FZdlkv9V&hP KthC WK甶Q1 Es ")/";}9/YbNfBhl ,y-|ǐRQ5Qݎ\d>2ZrzwQOwΦĤDݕ:=x.#9hTm`dM"THVN$ zPR`h2XACB _,|ށyz}yD xN7[ &slᶊ & j{UJFmvg^ VA5?n *t| 4z2ek,/'Q=˿'A%@í F,[ĉ *HF,%Dtyv(} rfP #U>ko^pgOY~sv@ 77Ga3҆ z8,Ug3p~ގ-9+k\ёFi0UZՕ@8g^ĵ38Ǫ'MAbڡZo1naZJvAQikWn`=1ϖܬxA z(+k!1M_: %ㆧbdt/>"_;A贫[Jx!CKY;|2>>#8 Dig?,!#豔 [8;7lz K@cRg"a.g.T(v*"i#QT}bviqvi8D]+MtXMFSjqT*GOwע9h958k2Qźvj~Z+КOka#=̺3m0Bʰ"6԰,d1/-< ]zw6)Q~ZuOk6N5{% I3)]pP yEEVQa|$ጪUL-Aj7ٸvssu,'c{yuܢZ_oޝ0yE1aٟ~9^B>d֩r+2ICgePD}DO[ëYiCST@7X"%w^m黯\(Au{_"@|=Bw=6\ 2€I)5:u]hJu 9#onu~dp=?_ZY0( ő0S ƊtprTbGlurfHNd:IӃt-tvsy%.F?*hH4^"ks'$OXHMzwF,!:P hK}n^wG825U7ަ ؖw|cJ[,Z&M س9 Yyg.ÃhpZID\@w}Fd~R11,)re29aR[rI81Ke,`*a<:D0z)9XG}P6w璚S+pQ}:~S$_,8!>w Lh'.9#<_vrf6cZ}).ŐY0WyCZPq).߾ɂbB#*"AFDK\I |us}M]_o+j6Q *AIA79Nj?+z训}k‰[K*˽]\tj+M B+RHdaqE%ڠ _c3F*tw (}Tm?^|6賺RVf L phtpZ-Fpۮ^~3v2 Wx_}JȆw g0;`\o׸V2g_TvU]xG S! jUXEÚS:/X1{*\J3uJPtN>UX#dR/#rbS.`._N:Wezk7WtL1:heIL!zҋcvLu8YY(~ ;犖b ?U۽r&-JXmmpIX}l\| LyebCm@R2Q[ tP(ղP ){p `X# L3Hg53,ycX^~&Bw^w8۝(ϗwf .$!x60&9m4fڤi۾2|VJ RbHX+f8RkI^0#)zEF CEѮD71T쳢Qh@Nq̻aU2UqfP HvPT:RfB (!bpsL½H%Kr׫ؕԴ/I6HIu}w-%.NUZV^*ZcG!<{48DnQHzYh8&< [Ϧ|̰]>`"?4FRHNO# UWC^vf]z= ^f]5ҶkAtt<^W #w:DEr0pR&WE~L)1*k;,T$*!ʍ:φ!| U0J! 7*.I ']noJWrKTӊ9-GSe N_\sl?Ͼwv{Q ^hLU^=5g| ta Oޥ&@gҒ4Fy85"GG2 gZQ$"*DQ@,Fl8kwLTN*x#ȁ ٓ}ztQRGB5:19DBC*D<'DbRRzM!nho^/H%3=0"L>5věc#sz{.Js˥%U:X%gPq>tq58BC-,XֿEAdk31a0?o s)N8lfC:G&M}3cat"p~6\g@\0ȅ\ jn|S8?Z%-2Hk7(R#T@`d@ c1bTYGNs`xǓˑ{ Nt>yffC#݃0 $f9G4hO3Rt*]My}{K(]A+R^"k]zTuQ* )E#%C"16EZtLidAmNy.3P"ptBll:Q#l:\,Cј<]t0v.ce6[Q<ip:ShhȰg5b㦓sBnsc،G7"1ɍHnEٍ) bϰMGrypEHH0LzWf&jL&D677|?H&7ĕ/j|Js<ڜccu6UM] 8JԎĭ g`E1#CVcׂKv5Fra1(PW *{G!j5 ;G8Bh%L'^pc`R0hh1{]w-Z}̚z{0T[F0AyQLax=ՍDp J/&cA%Gϵ0͸Jn297rl4[El[X/b4" (IgHMhAD^.AlfhQ̏ -LHS8(6EH%3N)!8eypR .`bIhRmϩ"c7S`}VV()ls6+ y`Spű""ʡI}1*lH3DP.RJ$Yc<Ƈ tTטaV(uz7.sOcqT☠[&1"{P@"N!@#BI '-_YNY=\<"i0^;df,\B扠)fba)VO^hp7 z?Y$z1lO6vgn1Ϸ-(Y%= 3/{7'uT-ǫ }'w(GZrٵ[M0XklnVrgΗ+h+ JZ+흂+x cM7=0!B>jOn-aCЉizFdzxTzwDPR/b"Va= LF?A%]4vG30 t j{au ō8h[pt;:˰P0٪![#ˢbǨWWM^9Ų3Se<]C`.gh_TMN\\=}&^0+&:ʟ>C^Făϐp<6[u$I.﮴싧N]rYj}幽SquHҢueUg)Ism/fP . L@ı(5d<2J\'}t)37 rW 5'!bQ 9)PKl%1b ^ۆR^wQ2l 7ӪN%L&T4_'ԓ<OG^֯8œ_R!-q0"!| J|`%ho0f:=KOߞ5 Xo%yq|T 0iI _g@x6 {+X7aGU%#SpE8Ukn(RN09)l}D$`;T) b3h{jmEgYq3[ᗵ`mF]*XZ#ԕ ׎>Y(pP>*1p# 3 [8.B{XST4KR-N jY\j s^̝ۋ+3 bC;;S?,NܡxxzQf9<=qew·L:}{* F(靈~!ޞ!t>[Εξ}~'x>-pCH>{toZ7TT9A=L8EtE7 "q%<>p:#Gd IjCqMwgUYeVuRz*yzdXc$GX! ][ua\ Zd+]j3zäbޟ:z5c暆:Ԫ-cMT;\ejTZ}z01*n2xc, qL2R6bj@_k %饢z(bFi6҂1i%$P6:ؽC ܼ_]o>MjqnG~(1xqJ,<-zbFoG}B#~v7W~4yW]-mt MOg9ҥIɸyZمOzj4ߥs;..8/'>u% )hKYO$kMGx@i:Gv8hJJ ]k/W΢xtWs]Re![C X}ڭfk8䕳h#R;n;jɗVe:Y;)Dgqޕ6#bb6~htiLmli_%. oP픔i1TjB+ ~ a+A3Յߕl~~s7n,爸|cq\Ηkqyou֯sP~)"] 6ĬVm<*0KDYe=V`KeOdԗuQC80,Wi2;>>a(&6Lײ]rʀi`oq;^^9eLq|ܑ2,[ژ0䜠rWԃUo +XkŇUE=j^>d ڣADgjŔT[zlxl 10ey5e l/M 5xZnssznsKR|9SA~2Nx^Ǐ_t\ZwDYi=( IFݽ f=1B{2(&=ks,2 \0UryٽE(L$ ~ &%*ߞnN3xf |./٪鳔iEujPs_|z~."VԼ@bBCAP^fLN;oD'UA%)=831%b7)Һ |.(<b|$ĒQ!TZU<Jֵ/0F~6} $L*X8#/c4V#*2Il8jFFk`,.u d#QM|ZERy9ˆR(DM"Ӓ[gisQHJ:go7_2颔3n H KRI '++` $ڠ,(@.jdT=6:,=3$n=O)56P4tBIu 4uŽ YÕ"X#19 R$ _[rXڏ5݀O cy3w_ \ 9 ^- rX\N[hHDEhXb|PH1&6_'モRRՍ=cNp%a`b s15|a%kկ[1ZCsC`U1D`F*L-Fќj,B(%ua ek  q2"((V %0G(vs f^VC VT3TжĄ0)//5\x֖^͔;L d^ +`+w)xe2@JPV^jMf0U x7 ooK AU]W q-5'Fƥ?]/%'tRʦPiZAN_i]*MqI:xJ8kfiMĦ|S#G[ޭ)mw;tnG#ޭ y&ZdS s6Ͽ}G]Aۜ6'9$U? `4 ~9SSM[r}10=hPRH2j (SDEe*otX,>A\TZ=KnI)ŃHf^]ZH{{yJ%R|NT}ӏ/WF (Ax&0+.dk@ zcdcT#ɉhI*4Mџ.y)Yc@ dN- 8>IzĶ靰pIJ,%˛Kod9.t|0|?_V{SY4WVPqťQ0G̟[cyoi XL(8S!/cm擅^`:(i_'.Bs7,(q'<&ՉoQOi4WozWh& K,-1}E`="V;A=K軘°j/+ͭz X&mYs l7ϼ43E^@]` G%_i0ˆ"u1BJÀ) 0ӋQ@ Oam3++R"{m1Ww'g8߶q[ C=bs"zm۰2goAT 3oB:-hsD(?d_]i4l'bc+Y2$ ZU+U BaJOSJjGK^>i_s9?â!XQ F+Qxn7ݍysQ6EōKl) koƁId 3#pTrIebЊ!B!^:"^Z\8yyI bF!(B^ d+ZA [Z,xW'_1}w2D2I#,n<֢c;S! N8=4WMLrjŁvO4E4U"Akw4N¢~xWwbr*¯x6҃g¿m"MՆ`*.#D3F )`7Ԛ6k&cW9|NYMҟZ1QһW¹H.*MyRge~]Ymcqp!ѰV `9^}%s,AuMJUTɫ{m <]&zc$Bi 1'_hayOֲ`'NV R&hOz 5OkkYB na$FcE}Z*XLalûe~\t;FkjFԅ[EhӒ㸚XUJ9 7>|WxXck>O&f$c*܈E5ؙfK*b൰J(TkeEplEX7os]dƠR(vs1ݵ(&6(G3MZyʣi4LK"49(flEZFsOquHE/ucOWR^nr aw2,ЮYΈ6_t)C:)Y@$LLy[xte=(RxS݃{yz3AГCjÒuA؉RCxpG3?_!;/[ AI#y18~@^!Aи)v_3 %%hČp?\WR輋Ww䁟o M2&Y;+9xV8/>k=ǽ3g gP? cu;OFX!ܠ0j 4pKC~70 2{nwPXV$u +RsΪ;t`~G.170*5>bFaܸA3kI^3`S+dIi܄bRE5}_WL)&=kC\|4^59߬5=} [vnMq k<ɍiSD>FclLiK˺C- 6 9:^[yU=8F/۟wR_./ݮֵȋѷpo=L+<~>;xl%?hNpx]O]GW`}%vu(zJcn{#7a ؃R!TzA[@[2Bb9nlmIU،vK(ɘA G>ˀaAQu0Qсɪ-\(EῐBk0H5!DÖ|T5E2%W meuU>h=el~?vUcw `htZbJ4!QArYebpIoDbGN(#G|>-8+T0,3вV(PM4&1&v,gv$"lrEQ$eXz{r3=_;-JWYWb&ك]^سUgi g%~w !R^:ZJJ &wu@>ڕT 1Y;9-9aRt[6cƋLc̹h<b*q"5B F ʃ1ոV;%g/5C-~]i)Z(KR>+ H0,$y-n?P47yx$dm W dwmI_!7r|! ,rl/_{3Cfbz jl7j_"0m/o=Lڳ-߄a%{en>ݐ!>3Ze9O-TÛY|pzi}vA:/-W#gM5l76e9~Bءr7Gl~DnƵϱ “dW.ڻEB^ؿ' ߏj#ˋ[w%*H]!~rqԑ|*S\Zhtυ71q.i)Fs|X%d0sٛhАo\E)J֜0vGT_G(!V>Z|9YXYų-{z_ 7̐WIͻrf&%]Ao?&|~2|˥J/ B\z ,sQ"K.դ}kU.؀zi33k'h8.,DEӍSAS,a=WCzo m)GJ$6.aRdviZ8%EBi`@,!VٞCPY/y:~^Qm s\j+(o˻g昱>%Kh)FC{/|sώ_#L$wy垀'H{dzON, cc*r3Z1ʆ$8baN{4ڥ*+UMD9 *^n\nRڂdYE)2t 1*C(Ej9Y\yhpVZV ¨D\a;:IGz䣠COMVcpms.iD(Y?.t˒^.ʜe8b"K + g1ԎI8+f8JQ*dG77Jʠj6'KJ*QruADvW^eUL 绬]߸JNBYu^hx KK f.qrLx`Aw' 1 {j6x0eRюcvNAx%,UTA.kRhJ*fKA eɈ4%o]gXƝjĪH=jcF3@ʵrJڮSl`>k!J릋HgvW0%kz O9xRg 6V6F0>tﯹHnHwS,[ƣo=ks6uPZ'Ylj FѫD.) oOOwcTc||FUyS}{l{]ry:mppӅ՗nY?uqEvpPhA]luwWj(;sҎ\+[d"M@; 5e1ENbxzp.Q6JƵ:AA֕!e/YWPPHjA=%-96&J #8QjwNzEA.Ql%&$rAB bǰ;M94D8PGKE`!M$ fueХ+%+F.vZ>VeOrf\aRIV4%$*S FR%y)nɘҺ,ɩسYvVLaqyg}|(*`|W͑1o❗ehI]X^J$Lm,ݜHsZ  R }ILi,fAbRnlE(vy\vGzqz;R:bXUZmCBK)0ҲpO("TV&bM#VJ|kNqaZ6SPmfe$;mxr K-ֺX\~mȊJO -Rp".*oګ,(}Tn> JcD=v5D(TYcЊ/)A!~zѵP(Z180TTB)X a]sfC1jY&d}Wd7!:s @H,;N]vR겔74.KyI"NtωPs:K*7ߜyX)9{\I(;a`78"Uv(]`d}f S-(սO/ʞ!1[?/B?<}y2Գ3قBV ZׁziEeHJythSJ y: 4{jQ!Jh~r0 D$rg>?gtvz **l WRS#tEC45R[?/k9ϊp r&S 5` vd]E}Y+{MddϺ܉l]YoI+^%? l`ityفᶱQ$%ERb,l4m#+bȇrf-{n1Syb\0.bLf=!,6|~u ?=Ezd;&5DwC(PYeET<'8ogwbz@[۷CPЉ,;hKP^2K昶>y&-ZGs1)4>llY+ J )D\JdY# CHZarz.C*bK#O-6Ok9 /gMr{Gran=`ÔtЃ]S|y)Vܲ<׹Z=崳Ts&{ô^ ו(7lwE8lТOݤyJacYp~\^,_M׼cԐQ#oέׇsG!E+D#ݘ0L$K"GH|O읟SDٚ^ zxߗ]}ӳ&',lqF֣+gջ`Kh%fk 7g-QDҼ(U oR \=|DcF3VrjH+xKWI0N1˶ZW\:8+00ikGFS=]Ҟ^>8[(2Z'߰:2ZΊ{X6!al Umuv/w0rLm}ةtiHM+o뇇j;{L/YSoi/]: 11CkI&KQb`KTS-׼cĐ[h-̑iVaIU&׳~qw_GsA} [0|WDjNSܽ#1|2WL$dq l2:^+fDj3zmyqy|O>Jl#ct_x;j\,ȸ 69 v͏;\{{v{vYnlnQIآɒ($ <o#fTl9!N1 M!"7#[[%\w'mM1 дvIw4L`~yD : C.w@K( }Q/fcY_qpiё^eby9u@c8FHGV\n.e~4VJuuhc%a-Q5똞r'~j _ŷ\p4(A6/unj9Y6hA`9tA )G}.z\NH_Z?]2/?$od*6oʗC?='Zt{2 w/ ؉ `4'*$<Wgӝ /YY/8N>zyeb1|&(i[﷓77W6RN>ݹI+uOK+Nm]/?h/^=qF/=xpA~ZzY|˧mi4+5Y+5Wk(\Q]s+})RD*+I#Ր-ZI'2 #w+d!hRX*E L^m|0-SA͛.DA`YI;gE(u9eVe`r^)#j UsmYvϲTkzg&y0p&j6R\6*r10&In2hSٲ6>j3j ]M,鰦3Pr-DVYNv0وBhIiENYDVTT7:saT `I";"D\"}l|.b+4?c"ʁ\TviWRd$(8l%h-DC?mQ;{ʓ#UiZhW| 8iCj&hZqeJ7I'S q}rNӾ On*?tIeWO,ng{||AR(?\p;}.Wq`3&{Dݜ_ɕxwD coV9wwլ{_//rVʴEaا'l-?V6{mosg>`9m$xrbiK >_$R6Zd +aelY71$I02vxmdT!8ݢAIS𢽒XdiX&_G_{Z%R0ƺTd'sVU<2] da"|D9<@}LII̓ARL#&9!zl>*6hAGI\<RҚcr}"O0b8s(ŎMi .'eidc'*TG['~-ғ\$2ft>&Np6Rz[MO*<8Oy,/H6lT8$Y#rkLHɶ/jlU\\uŦ<ۧV0DVӊ5 =\j`$YYYW70.?PFHƦpjX9!jGl*9"Ogrl]2L}oAsEL9mt".c2fbԆ5Ŵޣv7k\\Y9;ZZk 4X!/:-^aۑNnxY2N%Cx*&2Ac{K] yABOۓN?oCzo7.]-̏Kcp}R?Gio:/sޒL?Y0ͧ#EGcywTY> GqAГpB`BU+Ww̿MY|M' '_ K&O(uϊp>T}&;$;!l%;õm6EwؽBo=*Bja k }&h+:Atb<2˹7֑Tɘ ;VHH)YtkH-PFQQ`d#*a2DE[V&LCK*:/e/Ơo"z9=D@Gf7Yz霉Bz_jrl^c"t4l.L=ƾ}w Vc|'cFX&^N( FSSY5TY9M`hy/5d6PMFYFrN#z{ERGIe,~plBpcnhq}Y jQ E& ^GR&$,5eɲTxɝL CBr,Sĺms;*b`|v3ѕ"u>_ŷ''JT>YN7lH^ GہϗtuS)}\70Lمd&F,T,][s[7+,Leg\凩س3W<̦XJJ$!);i"\HFt\vd4kht+p{%_l=ՠ7hˇl_l59.o!.JKna;OqX? YW,ٸvPTzվ޵z2^Z\iw݃Z.L1{WQa)ZgN~|MBsDyvMLód5ĺࣷ:AI2IGJ !a;>|-i>.(5Ly" IC*Mze4$h*ֈ}tcsEV| l(Ijk&84;kx* "qK^YȆlܛȧ?}Y`s:9x\/itF8]:/l<6,nE.E~6O3$6>H"ZcjxD`tNRR-LDRM)Tdu*ghrR9L-zuɕ Xɘ .E&I3e!LAV B+"po.!sim 298J޷P9.ϯ429=J˥}Q8:: #R11:6kh׸|Ye'K|ew֒Lnq4V8v:ڈ0 Rve泜;Zsn$W_=.hq%^`t+.{!>@Fp|%_0 7{EcḐZft9)gyo㐥>@uRruR ohC{ qUd%7qȦRof,/ cBd)g;9Sʪ]g+={Y-bA aY{>lŶcIX"Y;ʽT=|R 5ܢ aE{>d #ָ{VָxwGa=m5BR{ň.oXO~5?Dw=]<ݻ\O^t{ň.݄'DtfRΧ*9^;e3 .#>)KhMNHu[vΉ #ņHFّyn[jeaEo#s9?-drbtƦ+/~Fx% [{4gpoƟǙIg{Oiw5+D/n74g5}MYʩ2(~|c tzuq{og{*DC,|^kU= Nh.Ln% 4o^' j+νi^' UxwJjOSYT2Z DqRqw`*5poUkPzi>g-9皶7CzRެ˱TC{s;,/P0٢4J(te?O/ HW[S:7ye|E.Zf0vyiDgszqY.TbM5mb[ۀTk[4V]ՙiZ"SPc6m7̧ۼ DKiB#Cho@XBہ0hQrq(O.YE9OzsGjiI.p*s&NR `G^Rd*SDXŬ:Y'ܣ) !%F|o}N G07d.'ro'B8n`Tt1:xa 6(fT4gHLtct[NA/C{z.&(iE;`nioBX2yeа+` 8n|`(pa9 Tԕ1](Ƴk29CnH4ҎU1G@KeA/k h.l`ZьgD*ml d\Ra520#%B?*񛌯+-n^ouaz;tȌ-g:"f2FF#.uA0Cm ETS ặ?\l})\h-gm#noI8a'>axT^{Hz+& 2ъ$_Q+fv]D8i}:{h ?z/˓.φ1~!UPMz/ t6VrtY vԸcWn*x&0f:;WY.lkv8qIɡredWmc0N@i$4%( O4倲!%3(8ZfBm׌{%bg18OQ[LD )l3#k()=>uЄ@Idk o@dֲ<$:Õ[wF[h69l@-n'n{M'+9YԀ4g=|~Vy`:7Y#G_D3F >g?|:>ݼ3~NgwZJ3gww|&p m n@|)ۃ6Fk@il|d;e,j$i ;j;eD(J&=RRct ",KԾ` *Bɻ sg(;K uCˣ{UTJQƶrR0(]c>*Cf6!cAY\,Gf{< ?>R0 c3Dk9DEG h4,pZvϖ_~lbz-4VR I S~vs8JQ x6#g߲x|[\@qJז/a?^m^BўǗ4N"Idy)A l(uhK䕦`KU1~j p䡀Z @M 6 @I,|+mb;e@vFCJA=k׺U+8 h( .#"@PPiyJw !3 \ kK~D!%?L)q'< Sc9*P9%,o-+$-N r2j#bx7'y-)moq$7$A ?tN^F>o~@Nۥ-& UN^(gtqڄ+F#;qڢ.:ʞl(;z Ju4^H7?{j\"Sjlks*`~^]/x?Ey>G!:?^_,&i!X4ʉEX4ʉEjbj) JQAQA42O Tncpwt2F9j3F9jT̀C}t:Fi4! iڙR#IPH! 5tFt\{,'iAE:~KMpj1> #%'cK# Xc Ֆ7eb *5:)jV7rN&|Zmϳj890θdbM>1N9v8r0ŮBK F4Yt#Aso׽* ^VOP9\%數MwHQK>p_n\ +GyA1C;Q~E Am5hؕ #<{D@!b,S>SdF:I0@gHT .ZnE]:R`j*n48\'qF4iF'\Ґ1P /OXPp*gm,MV,#wڤX͕]X1*"lB١H(!*5mٰ8|ͣG^ VB>٢0EJ&)# -l Eh4QAP"1 s%ugM4YuºԮUF>ZmlʊdcE]] F]aX<(ODԹ ٷQ§Qz5it &Bԋo׼צ&c;S 4gtLG׈'&r45"8DuG 2j@]ӟXsc4Y#@ { ][s6+*e+qLƳn6S3'"AVEV(Y$E"GzIb&htnFE96rBaՖû}`a8yLc%ek)-Nja (0@lV)T6u6'6&aOUse >j|&Eo$zc_O<8nu)uӻ {pcߋ'f`bܷ 쬴'ٜ(>EʕsIC.\E+TnrW<Bxa=:?;+ZesL-UkCC.\E+gHd*n&Czo:76]EsU`ocZ/ge=&lm[ZĮ*y '+g o3U~*ɌjL湙ż I? +@G*&C*s6ӱrl^R2rׯ)ŵa_.eXh5t *X f"ی26W11lXW,0#i9~>c{rJ'=}PeF@Ԇ =.`ꕭשeL:lNVb[=`{{[ -hAfLpv?2ί Y}5 \sZ8yא:nu{.op4Uafȇ[HYӍ/ktnP(;^*',qүFfxW٬%mhd>TBs̝C_#';@p mr0Brm@\kGhIW{F4 ͟}q-aTAH E #X_ ӏ$gsz({ȕ8f#W>}Չo|We 60n&8ά̈́ irGBN }}MSXcMMƽM0S|y.̯>ʣ?IY#NjH׶ WmT?ه 8-U'(I]Vz#䵁 Uavaz`! cow.g7Y.U\J]%]3g~i\0TS+kV kP׀̑Em%X"!ye$u`FuDzۛH$go|ޔ=fܻ_ A]M^)`BHAc䔎5wzw=OA&E}Nݕ~䤔c>L}BrSr*J* uӴ.y'c?W؏ ԛE{+IV4-}"/`<52!J<}#*Sy'7]-g>0-ja0{$Xo: d6}Jړ=Zt׿}gAtSvvK#nk" #Yv@QLHqPI7{*Jƃqy{Sj-(>)瑋NoɹeJXi5mOuPS!EP9A|9x1zv*@`Mڰ2PN04Hĉg?N2; uFbSU $1ݫ^XrF/閜ފ,: 00Hkww%3Wm 3WY3k/keE@ڒO{{,ٓ/oOA&19}B8iNY| 2eX3wH)!`MDmVO?g=^ ʩ w[wrU7,-ۿoV DT"(L2AP C"b DDQ"FB-5 rʟr*ׯ˹KqBDh CH&1!,qc!!KP0$FX\>F6|, yk-J  QPd FaY x#D1@Ґ9D$"|ڇ j 8aHN"B c"MA hF R5:q:4[)JX ߭4@nJ @"i ),*o$@IIn<c; qG71!vJFg9Q8Ļod]D9Hyo J숺 $ D"t.aA̫-3;5;?onoҘ@A )2P*RLPh 89C(JI"77F't\wTݴj휏,f'4RQ,DF0Ƒ'/i!(6N$&abliۑ?| P b3' c,$Ef}FE#pLA<Ԙ:N ;UU_%YOdu`,2N8W@(RkH4R BEQ#I7"2xC 0;wXDXƆ |2b` ^M޶ iL2E7q2 6~zή-7@͏Ad<oQ ]K׃!K IPn̕ +mP qz ƽMf_ 2Idz h}#|@@?~ ?Y;~>aZB+/1$w sAP}=@*A3hܵ;1D1  V֛*A%aw_PIGY~Q2=OqAySh@Y_f&?lX+k8D8mYBҾ\|Bѫ6*ʿrWmR\gZ,D-VB`r$vĺ [ĸn+bןՀ[BX~ lqPS}`ʘz~y D6tǹ WswF!sAo;xyx B3"pnZ8lVPxRc&IX[|`]} ʁ  ew.j==ZA@ bN9\kG,"r^+Ԇ f][(I@;)()8?|*Re.F8|\Nh-}Уɳ Cf'y*K6/L?hdns-F߾^|8oKIu.^0+q)O 'Eϡ S$'g1yT];EW6%9f,ҭ!^eYf䰱%#p[FGs=Bgl|2TyeWb 骫dlfH)$"4.9x׊g3lI$cɃogRZ)M@+rS w>+.CQRNmnP@(0Y(@Z&qt.+^2*uh5'OdT%w גq) 4V* j0`EoPo3$|$-'vψ*`M;ٴ9MF6;ei> Eb RIK$H8<>[ =.){lO8%*Votw+9G!%@Qi>\7:o`MzOF+ёiI~ڽ}SE:#3*hC]ٮv̝Q:]ѲdPԚΑhr n6ܟʉ@mm[h=yKY/kނ/q2$ ׭u'>lZ}TAjvON$@ *M*R1 v+2ޓ,f$HS"$TJfSցi7%CnwLp\B,aU;%>O͉$U—LwdqEzDfXQW!xP3#[{an?]×c=C6݇^2zVm KK RbxRf+_W@"9LԠ1N ^>|d`k\2,xFfM9 v]^zd¶=ߥA-' z"dqBa}'sl;?<t@m- ܁ulWGsNڋ8kfe;zi!5Yߛ"xQ}*ԕ~qE;\-e*.>-s!+P]kl q|\W@{Wl ؖ HL}_HCp^ nݹnARPYwF"pN4QFT{a}:;NGb]!\X e "G6P,3[iC[UuQĢ8Bl4sCUX$ jTlV34 v"44@E E&C(CqCQ84JlajO6wA|~ Aύ[ 3&M'צA ]6nWKz7wm;*I `#z>2 V_[\1 h0@ETOld޸*gUI`Sj>h8Pŕ([kGLnI*FhB$_~$ 5x} xĸ^uMP G54,J:B.Da$0 )`BBlv)򝣥~&/gb8씘dU*ˎ`&dc &@@T&Aj7[ 9.yv5M2q:91TpC4ًq!u?qkqc6_p#w%B36:Di)FW$/p(f0B𥝌Q_ ǡNHڱYNeraY1۞V4TakBKNY]\(oh`ּ׸+=H/-H٦ -4emږ^Wwk`] Q+=Uc.(d\ضB3MwqH_b32Y, Y;y~q`i%E3,_qf,!ub ]U^$jJ<+w6+0R0ZD.r.TktVl~,l= OҮ݉CNGAmz_6hx톓g^92;'O߇<.a٘|>$B=szC.UZc՛2<~z }ýIZ=CÅ4wG0ro{srW/> )*/)Z,on7/ϗD至7οY7.Ý] Z sWJd8NGL8'%mL{+}\''vSmAɵrXe7 ҰKz9+$Foັ;g:"KBy!NQ W0A 8ti!CRb6@PAdРpʠGMЋ ]RJ}ha>|#9i.ڒ^I1͋48_ًJ<}|n~naG߃FL`qN{X8lp <F07-J|4&:ϙR h N[)9g6(;Pbjl+gof{ ?dk/ Mtu2.~8al ~s>Āi&]W0K'>oC_ԾcNyx&_M$J5T ױlϝ&`2^_f3Fޞӿώk|3߬EI+z.U2r'VT1{nwtn;Rx89vh[ևsf~[se}wü˫rdMť(+fGw!|4/2_ʦ.eS^)KٔͲ)'^b4 kA2Y2|ֆ޹R\XmB3gOO秣Ef2rJ@틞yduXmZwei^w} ͕V~ݞF 4bh#VЈH=Ӹ(#MFɂ߼K]Eo$1*kcDL0`T. S Ѵ'K)h6)`q'oW zLwcr; 3GO_XHR2H1U9-&ePJmI*V$E19IaVfRJ.Ǟ[lnC%C-50FEo($w t[t41')\I)dT R 4DH=2ϤȺ$uZV)3[O0ɍcG<ic3Lx co3G"}VAiYy'rc\xi'&h2lM&o}iQmi$ʁI/9q@)M`4̙&4:gF2|! M.^e`PcFp*(fk {w)\I%$kzj$ΓS!Xߜqgnw X¾ S8she[$|V 6 4>ηĮ(^^/tzN\$O)fȳ6.h.G7#4\V;=۟:OH:bB.) ^hK/fV4rJ292d+O}c'/BCIvf=#aIl9B5nW.Ȗ\iU2U=So5*ڴOoܷw|)}톯8`E-"a \;KÓixrc[;8|6ġK7wO [C9zN] wv#ѢS4Z+GRkāz\O6}bQwի6ФdR5eԕ, %pgh5$&=淶`D}d>5CAs8y#L+1O6*i1gM+d |>kX}r] icVN{:4 {۞"f[2Yi+II%[lRX.ޗj_}L bЦvg;cxvW|>#ײlEJ̶ k6`Gk"V?4\o!G:6Q96h'S{I!k3T#e)_;E z-Sm)ѪRy KF15wLQGNrĥVRJ^@TɨyT-~ɮZ঒Y1ڼ4F:xycs+1h #*HrȀW ޸1>/\L̀(%gk̴,3V5 lW!/NJzsCǔ2Oչ;bv8iiMםQιGIU;m *J-~k!kx 1 2gܺ3 E6#2Z)qThA <%Il.35e22 w24WǴI_(X Z6ˌ,X 'YJs@DY6Dq31FTwd;AxC*!3|B]F[V=/VFr{ VByRJh7wd9odnfd&ѯŽw<۷Wn,!sE#& 7=kI D`NSpo ,3r6 ; l@056Vڸ66V2#t#Z?ƹ@a:6hȀnI.6.TjXWGN423ۣ#KG£?c4]ݾ(WR#FQՑQ!Y0V$-LX^rrqJ[k'Jru=jYK^;[O$Js]-#L}QV,-U0VpYEU`B,F݇m\_ݸrv(kO"x;ۭ'5e@ g. Z4JR9̛[댢NakqM6ٜ ~4./ s*|-U͎qZY͛Y<:7$_jdÚc;tKdƄ B8TӉ12kd[[%K&G\ӶrRPN9̥;)A#p`7Ȃ+r|ʻS) jR6ʴrzE:\U]ylO":EdYc2N:9e??XNyoNaiYpB8zʰҗOJ9Jr6cv{;HqAko IxCdL-i(tC=2**Al<a.UQSR)*45EKjjw *XM+&jcyK( L`) ,MMES10}},D%dV7 jR6S6 @oШۢӢc~2v7;_JINT=zbBnzzLyyn! x+4V˾kŒ] ; zB8cieN\.RY f~Po+=|BsT'ҐRHW\q l#SJ9ҘRq8ET1J{ioPaC*-omOF6]+6] WMDW2aw+&Tf/|:>7p8Fca=Qs%O RaHӂ..YlHLq[6`)[AE\ ۫Ni3/L*n!$ŁwB҆y㉐+Ħwh܇hʫ2 h1qlMY ?|2a ҇!)ә/b!2!M:觅V>= #GZ=phu+gWBr$<ᄉ ><=.4jʎEAƮUd+ݷdrlcA)AHNkPfrP}''^x' EΛ+bBD[A{ &᰷WLn0xiis- 'B\^|R7+~^2ApŹIDጲwFhkEF$,WY w O\>SV A &Ld]OW]wOܲi}o4B>t3|zj ML[w~Am?1_=$'~?ӁH2x=0`I6&m;כ`rRcѱ(s2zwWL Ȣ*8cgC]3vRA'wo<5PZoP(y+^V./6uFwct@  3UpL@Ys|f(_{~e6܈`Qլ\.WRޢX*xcb<C-\ QYy˖蔂^UӔdrEX_c R% ~XN'4HoQʺ%z%Y^ɅnzJ%h92WOу@,WX3! ̯u]VJ+Ą3nzJ%xZ9C\%MGLS ũd(oc "8F,2P$cEIy^­%&dLJ#7!0 {Ո^*AW_m+c)LYgƨnӓGJo¬teLRV+kNaeTҨG(5GrOF=Մf=KjƧӭw%1ұ(ڤc0E|ͥ5{j*Կf߻%eyߟu`¬Y\n*wZy}FRdR&.LsQ\TRa3'}80fCdD,CGOP=Q~|t`k` @Jޑ76ZQ,1Ee5%ӪpUkaQBRY0`%RGp1R''rv)ENFG=Ԛz7뙒jqżۅ 5ҥdѸ/n.=ܬeIOx?O_٬9X $] $G|OשH$`Xo΁i`fQ>zl䏛\)9G{Unx2Y37$FݸG7얊A餾#fvP(xg䎆n9,䙛hMyTM{gT BL'1ۈ{Qn rX37$OݓInN;bfH@^gt-SZ+DA29 )%E Eq*3JK$*lTN

Z)J !Бl(PɆJ'JWJ/y#:CzEjrtP2b^1 J!8X*ƞ~ui{,&8=iů;fA=ŸO5{c a tv_&-U,3/+kW| qt%R͗K,5EU2 ")lbDQD$n"\LOOcՔbDy4?pN,-jw)5Ňv4KFieB+&p%#DT6/KU0adaduma"ǪI:f[<3udxw?\} w@[6SE#6hA,WڨѰut2C [Mp$%td/t_stC]L"ZM[s^!x>4+X{q0էww06)t[)@crOHݝJ ϷsiS]fٗ1GD9zHԷrıW𧗍-)z1 ƛO;b^ Bk4r,g"_ʕ0j[Kcfk|^>]^΍re'|Pw]@ݹ ;-pS0\5#\Q%)'e] %%Sb4u%qUsT~ryF5g_g?lp&mj/lt(~oxpgoo[oe <-}ϬXK=y{^PT7DےZjk4ҬJ !]JcTUiY9-7%:n֖Ŵ=h37*߳-a u8;=,H3ɽ; ^W>}DDpsC Qjk0>]^yvbgw.OClQOM6Ă2A(x q~Y~oǖFB$FaL@y,P,vpbboڂ[no?vƼ}pVϣc?g'>Wկ\O_jk$3w>b·Jxy:Vi·yM$R`4>\C6N! t K;LXRcbD-JjmADBӪ%`JjC:*&.`Ԁ{4`R VBiɅk_kDb:-Ew @[?.ujKǔʂӎ%"50hFU4ԔB A=u@)|'4~ҜNThxÞ6ep X4V\_7 ~ۿI Ȉ%*<~B/uvXw{;#kkcN-|W;kRׯ^%7}E;@`{AvLmNf'0׭7& |&秈%'$c*8B2lRH=|B]X$P ШHၸ@;N6@Sd_\1*ddoR*D;2kKqؤfXfQi<&oWL̲# Z0*P_[K-*o0Ԋ2,kػ#˚ZHJ7B˒w$9~$ۃ $:9}]>'#Gc ϗOV9/gmc;|~U1N]-۱N>φHޟ s=.ϛ-(75z77'|=gI`D#ͼ`8Ј:FвGŁq}ݻT*^lݥVZzOl{HK' ÂU{(F:">&REI:ooO2)8S 3-Hpӗ肈0Bȸ7yn[HP+mL65P̜kBYT Bj#߀:ݷD`eҽ~9,䙛hMI,^YF62Xqq&^tvGPw9(yb= LqX ҍ|{ Z_ۤ梸>oUP )-6q:Fffp(?m-G&/6M]Pu7.`+JB)#FKJhD)G)2ev }t+%3!-q uHPIi2:qI,TL 9 \r55e U4e]~TJ64$U,{e"/L Gv"biG@hS'/ukoΝ;{;"\keYՄs!DntJR.)!HSK9|/$~ԗw6`hИ9'(b \bU[FJ p +1Z4RH'1GFjmi1嗊otFFk_I>p = LV0mqsY:Q2E3A@&*8Sa:Dky(qX{(||*"LTMЩN݋ħD<[՘ ,ѡ} P(3ul{}h]g?'t<_$/3ހrE{(1wI '<*]@j!8AQd%np{MRME$D?{WǑ b̌Blbذg05I$%bF6jv7ut5ن!Rb2ȸ22»B?U))@ d radLy-_K}O;#In#l؉`KУupP]{Ajbq yi6u{~}qil'i5KhFr4{Vݐj[GlcGOK.~TK?׍_jHՕ[ƭ[[K[kջ$NfPAۄM‌kt8w {q~zAzqv䨭7P#.F|r/ܿyc7)57=l' Sv;GvۉT{,XU)6dn MBmC-nM6iqQ{@OvG>#&q˜ԭ+]_1S4HP;Ϯ+TUpҩ[>TV]ot nwuw=TKbb:^7m+.Zn5ZtmKӇ4ŐmFhmqZ5u aW ص!O{֞0ы a(p}w7wv_k`y5W(2 4c>T*#S%eGJ`r&C2w {g3.5} F9tAM{kt頿r1YW+;!!o\D{T˗ksE [_ jD]مIck]*ELqfijԈN5hݝ.tӏ7.nCH^2 UZ=vbP#:}v#u3v_T݆7.[˔RS8Xn1yS2 H 1ȫ)ډ֍皇m"sp*%}sbwCТPQŅ k̞M8n. Ui7OZ9L 5Ѡ>8-3HysJZj=Ѡ2c;8۟;o65E ;𔶃#Tc*B>0 a| j.s_k! Xy~!xG8DOK5DIxQ[ҡC{_OgGV<욐?wכG9Ҍvņ?)?^}zG;ݗOW䧟B_SbރVnb`{ҁ;CkXbVkvۏ  !aZ3qDB%Yp¸4LL ƃ%]PXTfc>9rHynJl=@%"00ڜH!fXhrYI:ݮbp' *2h' M8;wjcF6n[O7ՋHԣ`i7FQmKf{'b i}{in,+' lԻ>*!5bɉg3,+b (DCI<hQV6:^Jyr^뗣W_~Y|_H 8_N}sWAn_PتZ6`89n!Ʃ䛔< {v[5O_褼;ڃz*GPcHa3*CyO{&)AͯEqC'8ZCd蒁!-/Ǐ9#?dk0}o#F ㈙x}lڲ `w_oF`%j[ EʈlFVVJyUI:aڨ|#䟧sy|l[#$G2L("0sX$0[Y mrxQLa _Hg/H]ebRdgi$oBEI\1u:3:2a䀘s<o*8i]VI\d-.(6ݴ)RL84ad9(  f=NJ}Jh5e?U=DJɤ;,g.MhA{aEh@B*K+ڣ V2WVK#m@B%Bs`;(IJkbɦcQ:!R}L+>d52b5Nk3EnPIFYu KfB>Z9yLef%}y| -Z Epr&H?npG욐YߟV_Ǐ(qx“@DnAQ| )Zw^񈄋NdZ`P4h.ď|9S ={Ni.ObY[xzcOx#Vv2JusPyNSRF33H2My!!hn``] ZX4A" tG㤱A';!=ɫ):LJAHN7zB^Yl*Vd3t֘}0!B Qٳ#'eN` 4YN'ihp7Ҷ &+ BtY\$K}D UXdQhOɌڬ~AL`|\dW~X:_~"\"Y1?EGm +?G‚,?;n}zG9Hŗcno_Py׿|wL!?~=?'N駼Y9w7sjf?}FlQ {"ɮ)>>݄_~}LkjƢ ?Lܩ hot>Δ!L4ۿJGl!5bɉg3,+!!Ơ5jJdƐ.^{y2%}0rrs%n@,Y]I|w|==J.k]_}uyL{O.O*=Du]~9ga RB9' jpݠB?P ~? u]R~r?;;}+2`|)/-ޕE=>? ᯍ'G2eBfOrx|,ۇwO7F"&.Q8k0k.-ztCHRH?9[ C/^dp޳{BJ߅D$+7_. شX-YytKz9"X#\;-oä>cd0DJ]7*C x*kx" du6&zf]ْqCTpED+5I43d0cFNx5ixilHgq~r(W8;ʽjE@yg3V MBJҟPpӁ#W'FV~VF":0Njp\++O}vz8E l:P٢U(H?sx}Np 1^LhzI)ocASy[~5r(HEt0(E :2BYb^;s+𽓎G7dSI3{G)m1%r/ψ+bz$kh;21Rm$/[=~ʟjnϩ8 ;RHRHP${PZA+QR&z Z_&(>K9BمS5!"_2к vR(L%Dj [_`VJxk_*#^t3xu%jyaߴj.oKhM^^  xqV0.Ѱ41UuA"PY=[R)51I7 .n{[C^@ U&lf#5V A$Z5A#U+g$1Y U' ~-d NQ'!HZъ\݀} К>Rto6 a=,jqS9$jR5 MxSI=o Ѡ9.VyD9#ʹOpRԱ)L*QHZ)1`^lP`C;# Xl/<[})$ ƟUjK;@6{ VjPYz7?dm&T5k ̎Y.wvv}X3Woonݦz?@~I/osfV u aBo-,#j=!i˽s;> y%^7{ysiL@SnyPNlZM$䅋hL)43v)hT bD'u6~ ֣itgvCB^xU1u\ W*<_u|^@׊N-̟gLbpV8Zpk}֣r{:͝kQ%]>* ΫUi~p$!0JOBrFwuW5dao5YuqJ)=Ѳ Q&T)'2#JfI1<^zڌ{2J+l9ӫQ q4I5C3{Z $̻OiAh ͣncjg+t1m v5D! >m8CIX2Rm :DãKy iB"bad0ZɱSr8wy>w.5wg}E|j}E=b} {Kξ7ؓE 5/2T c~;Ts)ǡϡ|$~V&ycM. W}1-4(&"}SA[w sB7ʼnq{|a|I2b 1<=:?TG7oa;zruxD+R`*єOk4?ƥČMb\#2 Mթ:TkɨAQGcr"ݣgCbBQa\͞K%LkUEs2@wa>8 Uwo7kf;P5ϛ W4xgi"*Jju8hN7==>3ERKt9/@͟dԕO!uSH]RW>USWoͨ!y-43K6Io O-6zM9L[Ti*k0vq;v21k5];K.Z.s\3 4e:ܛ{\2sN W0;qa6wha΍בCU U ڲ':J@f708ZvDavM*Nc${ƞ %xX"/ PYN5Z;+Ix.$>7Q">{W)iI-MZ~FP,(r` OSO*+SN' cUC q +ȩނ#2 f^\pĝr܊BI>Wb5lq/ aH  nY rrA.$Ұ81=/-+ySɍm7A*ORyebx~J$KAъ iH5΃, PT"j$l?4X 58:8K9VĘ>ncUpJ(&E(` IxOAXa1**4<.qÈ>?f)q-*}ZIVD8V_  '0#G^y%ohAbBYIkw2%&0[ráT+{VS Y`n 42 PѰ=0պ9#k*Al t%.RHR]v!Mpp`}6Uly4fq.mH̀:A`v'V=6=L*}Kg2-W3vx.ͥ>t/3.km\ر=EL39؝jMOz sGhTj3rM 'ZmZ/(xZ ݶ/ym.I2յQcvK FtRhoJs[Dc[E4I*:nLU Qb#:cn YV;Ckj>$䅋hLi'?nWڜQb#:cngtNM|n}H $/'N"jT?WZɝL j;OyNIb;r1ˌ9 syƤS.wWs*%Dc^Xۡ(  ?]S;VΈş tDkw|M.xU,P cLd͵k(OEUo3P~Yӟӷ"ջ gBoj x>XG`d ͻ?cD%#r!;^"B!!u5s>J8'pNl)zzHgDZ"_%wSDڣ>OSU;lt f߸k8FãpHcƋvWӣ3-8*yut.X) `!: mBM5&N7dZ7%!\ 4X-L8 IQ.Go{ lhUOFEJ:,bu~L%Vg,tݰT:$DJɅN*B]%L'I9 -Ц2T)x.s:e<͉GNʫXi"=+b)ω./*挥{^8cF,H. 'Ƈ~b֒\6јK$'tzғT\q*Z!-NUK@z~{敓O)+RfkSV 9{̋x . 1& 9xD*[~=؞ln:ٖ>uM1Tۆ= ЇG17Oz2csc$ɫvohUu{ZpHՃwL mB(B-ݖvS' >.6(gz[_;3+̀sN L%Acʗd[deKVjĶZWbXѸSUp;8&P3R`->zK06N POU*J$&UBh偫 u8缔uQ9^=ju)>C Wi ^M^S~kAzp՜|ͱ&@ ۵{d#=zHO,̚Fǔv A~#xEuf"JYҍY&o۞d/Fe D* d 5%\$PdGzI-%6s<ڰaL,[/J[eRٚ,9MP(O VS(uDK26#\-9g}hl:g/B~ؖ[}30;G^47D)k-j_(DTL?:"JljF7?CѧgAʾ7B!{;<ُf WeMGMd|aPgH:\/o'A?\՗kBVVn'YJ 7_j=J՚)2rl`jYRs%ϙ!:*GIB,:^ "xAE.dY!巓6cŎT)y?{*5U ˁ-_nTeR4HIhmikZpuZ,LFS+(gumNpnQU\rxN;vC ߴ MA7F}U.^h@=l0OG5ܜ0/5m'v]zhn%crbR2xH}ꙸ_yd Nn.a4Z<ӕ]؟NʁRg8NLeEf8->É B6ָ gM^==|Qghs@TO4PZV#8F#cIK{+>MOvt7u=uV{Xd0>*JɳQO8/%aQFcMqSMSXSDShIJբ42JY9rb\ƅWk{5Nia29N|vR-08vVlIև_{o(^QӒܿKFS#oF)І(35@k=f?m{kXY`]PsxeBJ5 q҆J{cӡ4 7 R96ySK@n u/ʭ[EJQFkzW/q@2Mđ4坔j'D(AՏXt~+Eqnp;]>8Hk W;9_Чi:.=yN(W,L)7nE`BzZG-toVV>vZQ,@)Q-xKM(dau`F@ \h^hT Ppirߏ7Ӯn"'wvJ3=UhĆhƴQ)JWigRx˅B - &+ZVOc('((HE!MV0#Tڦ`5ßP#έǡ=ftr?>DUlؔw&üt^yWvÝ&;,9;ҰQl:lk[C)`QqsP=u[2 `|Q >㬒--s,Q8FcV)k<e1r<(ɝE\p|6iŦ{9=(!UDM7e!(0$ Y./O\_q;_un{O7vZffA:eCp}#vl1%˝Y8RwǑӖ$f`8; z[*^* GJo_ %l4kQP -c ;swJ^кGϮ";* jZ߈wͳ {EtoڤN5@VoQ^NѺ-n0ҋ֌n7He;!ʠ b'+6*e\oM}OX<u,x&SCgB7NGO}VkZ|SFK[C ޅ?!>Ô+NG×ˠQcb+#@DpǕTQf- LZPM ""р?F Y,̵n“CLA4p5: "`e<Y$xT4( ۑ5 (9bZ5ݓyAuܪ㫂[:\Z3Z9m )P%@pFD)-CYi"L t;؇ҦJ0:+d*wEB3MliTVa jr6\]l/T |ReӃ~ac9>w8g'43OLJχbNfW|@G8MVcg? j2-˦U5_}yp}J8 Nv:vِ&$hx>}"*Ton]i^ޗZU6gN$ڟ(ڙKn|)Cɐ1-D)5ۦ[6K[k#fi+=PCNiOhlߩ.6)z(ڠ`G7T !6? PBi)p,G4 }3m/}JKO>)d:בCl'#D|ND|N ~ru;q"1l,Od}@dЮ"*9p5*Ճ|>ϣgO!łӹSB`np;_>8H=jfrtT6t9vAhxr_|741yAkxКK5҅ңB-+vLO듧*=˃H"i.ʚ||ax AiJ8erTk5_Q"$_%w˗^0d㪓ƅJwWeK%P %E$p=mEEDɳZqC,Ss)aJhU Ů!H\r$J, ND$ `jDy3 +}`cLpG T ZErŀFA\pS<fC4S/L&`pt A1m&qMmOsJ<]db77 |-P,ȉTp΍wK8nz)Pos kţ x5E[^4B3j6V0IimyyqK8x4ƪ c)ph5wkR-2):'SIBJvf[vйT+*o..)F%Pܬ4תy~~Qb''J\"3Y >>sȒd)?? rL3T)fcIǹ'x~N#0s? gw R,OPӟgfQEᷥ?^R UF+j:3\ r3PrW證4ݑȗ_,#6 ZPD#F ^fZfL* B%4o%;aa7^82fyQlE[>S ncJ3P(N!Xv|D LA #j\S\Jwr" Qu΋N`|׫!,)GO׹iNg\j\+Qa,)CP"?S]ZVתnfu%<+ Y%.WJ NjRwޡQ3א~˸ZsDycnNIo&}]0Ʒ:DfDcv X+@U:#vɑډ<]Lb;>t-p*7mA3mɃ:Q'6n;?xK:g[eݻiD$7soHY\2kcskkk됵a2X(XKE,$Z$2J2TX*hAA)Tʩ%(v~v tvO+g}IhmC+PȓU"޽MZ=Zk±1AgnXz>jU]&cfC2D: c%*\#`-I_x=KCRJxnj[#_\롯Q+ oIQ~J=NU QONCN08U@G`=r` ^ O^%4ͬv-JSmAxQ6HൕJJ)͘ D ҈RT҂ #k.8 JDzM7Px;Zo)P_Ͼ<]=;sׇfmX,.w-gk4.& ʉXT{9+aR*:(.r="Me4N* $5r % =N;K;DɲslWΝc;=CmW$(8rYv[yA-tJ^biPqxBP.$0Zejs#4@ $vPѼά7qJ !!aT"}Ycڷ zjVV`+Ag+tey[.8\S8d{.\px<#CBس܁ \ 2.Z[rqKNUH4e䑗>2v\ Ld~KӁ Љ* Rj5Z 夸M (Ĩ"b2*.WSxئKu3VdhRFO+fzG08pjK\%٤YI NAfOl4^֖BDB2h)KAihL.Zk`$[~hpfosϳ/f>3ғg2CoHɛoL~˩b8 4?)Z 柳j}L>:d&5Y7Z%MhʧhœBw ]™JӇg49I?\9+uoa}<{g-@Ӵ )sdiiSNFѻr9o&wuA h 3 j ĞYO/8&8qcz4Zca$q= z|q'  I‚4z$E莣l`؍b: SOOlG?Tà4qƖIs읒G28[C2IqmV $$*sSjEhukX3yi]>g@s_9L*Y[5 vWD8Cg¿ډ}_$-dF汓uvZ7Hl1bF /ሃ!;AKDЪ JB" 6;ܯw@ǴUGJV|G/GQ,n煗1ray~dqEp<7;dZ#xed=:mGH:9EA"^g\xq^1ܽ؁N( ^i )A5w+5f $)Ԙ1ԘmX(`@Mj>95f< Pe5JP;S+&Ֆbg(P&r)ŬZ=L]b% TRnj Ԩ )! J3YDIP@08DO.] i;*'sÁ1~*@q$"b+ %TFä%E]YAMXR XDU@5$PԅNEXC\DSduj+*dEvK16haz]{ p*TukYNH [ע0Ph] cKi"pdJ,@;-WD%aԽV̨ T%Œp+F$䈯Uo׋Cd;Q[#tH|_~#fwN~Yݽ{i"pF?™DbɃnoǓSX?#?wn⻥)&'pmnDNpLVݽ\ $@N\4]ḭQ )r] aڈn#bi4/AS^:QN֢Nkm) ",Ȝՠ16>NNdNAZ6& łtQRLodJ ca%s"LhJF{U(]ZhdtOH_fPku2+ 5+uE^Kgp>xX pÑUc~BZ;~XW@V8A(rj&9m?iʑE9 =1Ͻ>/x-ƭ\V+up"}Om"NBDmQ3J)ȩPAi-0:f`g}XF.Pvy Uw9Aq "0UO.ޣB[ 2Icqd&LJML҄,J") HY8ʼn d_ %ՈhHFE%<dfc&u c4 1^f@ِX{@USyFRXMhiʌ9dQbn3#mEUQǚF57JifXc됪!nTul|n@ʚ |.>n< U:q,~v.Q_3WwUskxg+~c>a1 [lʟ~][&gnºo߳[(+qg?+'&:2ڎwG%aLw;FidiQ򢢕}P0VKRmFX 쬹T/l#cmMi#Vaf+Fz|-a93;`֘ѹ|mR9nJܾT纣8i)u8\bb 9Ű?{[%è˖M٩L6yR}WXB$j([sbQV>j&ĀƝDž??-VַZzvK?;܎dqfnKƅ$n-K\okԢ SRf`gK=LLأ5϶,y m?S\Zۤ0F/?pL+)*67˂$Op\XTGSh*tzg㈳+0v rd?P@h ۦP\uu+4c+)%C[K 홌NL=1FZmT._f@l0{ᾓqxb,mT<: MŤ `E\~囕Mn6[bk 2t/֜k-:߉Z?9q7;NމZNT;+?unx ^nxR!K3-v:VRi!UiK3UIK% i~ghOp8ʖG&(_ز4 H`&ULIdi*<֐1d$1`))4-KU)κLeLՎLg^RvaieQCM*\%iuPr}8ep||ZPOfůQEXhU=ҋdGM?h2xd: IJ$ZafS?gIkDc06p+ ($JPe<*l9\ y~ hQ6prMG4NyF䂿Ofq 1d2/Qo#3#Sc׿I9P\)N~YBt[7 pmPA{._}'#/el>̎>.6ޣw<(Z"~(KAZ%B-\mOT}hjC<%R..7>ܹ1Kp+qT(nexwY MC; T:=~|GҔސƩ,H~=B)@ vZܡqAGD bixqBhuљ '쏷)]|Q6wVSZ[<[ܖo=uvEmc,tSOz1B,('v߽` ~߽ii+Dɀ' A!oCє5VnkQ T%ڝEk+Xm.eɬTyYe/1w?F%.W͟Dv+neح}S޴>.DR$X 2s i. BJgBU2Yl0fąf|^34tlcE׳e 6\==ڝ t|ê|l9?$6Yd/n)yDytpS#GN&5.Nx;``ˆ}-s߻r9#ޱ]QY=)1;\w5Elavc WLs&ry5z +g̜67s۰b( o>@0FILn#QW(PBFaC^1 @O0Qx09"]C+$bYNzg)wU\g5Ýu_L,<c6`JvYp gmCn6D36%65$Uㅇ 0+8Q/ +qCG2]FVT]N7d߽RR"4K)ATB2x 28* Ʃ$0H OLGDozHa<$H5ZURFA<U?gexZkVZh*K3.JE&FL00 ]2OhT -e+96Pi3bq B %FSl0(" :yI8 rf 4=SQ9m5{0."r֑.=Í%?OY=$ v9Y_7Ygzɭ|g[i) ڲ>uU5UÏ8ُw?Jmp\6SY)}cT?#xi~_y+`lrIQkoc<9'hW ^΍nam3.=F&~8~^~TZ#}s d؜YqY#_=>6%}YTWe?_nuu{s9oo ,-z0[/|\-6}ɗ0@]:Jl'Mfgv9kiw>H3馬RsL9S/SF>嫊*k c!DlJMQ>拁w1>Cj=w/BX+7Q|7!jGw+_>RkN뷲'EVrɸ (5Y9 DȠ`ڝyV!˛bg* (2tIP4bFDZ Lgj:UF5kcdfF(zRI4. wYNl-a!DlJŭa:d%Qe^ITQ^ )pWₑ7r وphA wQ r)\?! Lx^ՂiC{j;QI^duƎ'ak!fJ/S#vz~H&f^r Iq[Kx{TܘN0HjBݞS(dPYJIƣ8gI$"seIWz %KujgV3þHARFD8Ԉ()`? #9 ;0 .5Zr3A2ICWro#2IܵSdN3 ~)N:8:m}.!qg kapg;.q)jʓc49Ll*6DNHS91- , ;GJu_6X5 -wYоŋ3oYE_ؿ f [~ Dֈ~RfYPʆ2YcC=kku,7lW };3XndO@@RNf2Opd)Mf A#$-,la Ui jd ~ٻiR.t-݂^"5U=#Ӿ*JWMNj!(H*GN#qVrJb)$uP EU%j7cf1i5%y//zMfx~g_jE/qrw?FI\ɪN>7v򹱓ύ|nϦq%%bb\Ka2j$*e)!o?K8Z|_t=ۤX "pa.2w6Z83j>xeJ~n2K[hgj{_l|gqJrbw_`VVc{ly#[ՖZ2[֋K,>U,S)E6"qu ue]׿Np&)?NNfU<_]4f-~Gm9-^*F8!!}4$AS'jrB*(- l\a^jUܫ.ks!,v}5(8vqUD%3 &5Aj$猼\qKnggb49땱VL9! $syP'1<11+Lf]ksMV'(kq&&<pdK1oήu.;""Yi@a!$^6suYameB = cuSV$Q I}X?&3-H8IGY4äsb=n&-Qv{⩎HTsViyk0%(Z']orp~.Gdngw~ѧ!,,Hꍟ/R/q>]ۏIq&Z+v=ҿI ~(C {h'76(~+5YwF3;MS̀T\ζ\xoč(Wxiȕ!; 8@1tz @띷suxf;;ƣ?ܬʱIjr?2q5JQkυU s9MۛMX6vO>t4N4.baq E^؉Rp`{!SX{*ovZ )ȞrS M9n[<KeyM6hdPbcv+/x!Àdc3LR{Zŵmb"-r֏)'pEs5\M׫9tWQYg1feZ(:`VD< lszj}Γ3E;鈿<9 =@)'OQ1{^:,l{H4?ptu! |O+-x_";9!x2*fjhʒa3|˜"͊c;R ’x$622eq++*mJyv.hJ_A 8JYA(onG(YVZ!01<w N懫>VT . D([DHPCP 2D{E˓A5.V.YPGGk4NQ˰gZpk /STP,*"K֯St 7Iz]6I#\qшY}Np".2Wē76= )L1eA.1Hɘ|h%8T +0$klcSA01q`[o("ڦw6_ mJ(+' y_ZWԀƑk㠎3H:8AP1Gu AytܓHԟxUP|d@Zg)O)VbUT\JdB%QA 崰^x`"3~%bb(gmsi+t_nbpw i]O¬Zr~x{aI4O<`_'jyN^{zP&z:ٻ//)e=>>!nCenݷ"ohmmSM! W )Sit z$ @k֖_zuO dW1SN HԘibb.HU/ |dDMGsМ/J ݡL0IF7Eq1(s>fL:`VɴHS I8YkޥRI.X[!KIiD 679-$R[ˈ3 İS)kmR3$޴No]JN>Vv.r~E߮v|.K d[k% }w2p}n1}zg7F j8sNs?O{5ƏgC鯶KѨ 7qԐK2*s9ʬns*"Csa}7_,Xp^/8/>yz3x.:i Kydz7jk޵Ryfv_ yf_]b3Arʆ; |mJ_ JVCͺZrZ[aGpn[>ݻ[dط  n|Jgw괽b;Rud18oŘH8=>Eb=va/G؃@ʽCD!ܑ{6-o1۴㿰N+8Db-*N(Sđ;e=V*bc qYs!+ZGILƠB%:'@.5v8ywRo(œ@ꨒ o3DtL?=b6o7Ce]IAv{(%=. ':.z .P*h>/RCPnⷨ_N^)~I 1u[ /wz6>cad= QO'\M'Tzt,ZaqwW""x0fC˛[ɫ>.$ aEvVC"wUKPP~=ÿ99R$ERJ, w*G8=pMyK(G_Jݕ2[?z3vHR¾$:}ˌ_"|3_~3ӸWPrDŽr#Eiz>֛ԟפnN%&&F"C Qb%Y8s”1bOG#l}pWĠBǟ憛y?(c.yЇԸG܎ËSYKj:b~ɗcSC͜zy D9z3}U|/U9gKBs-)ɏ>bݻ)|ޭ)w;`M)C{nԅޭ MȦv͂c Ԛwäqy@VAŻ0-qz4w+On}XwnE6k/F׽[)9SGvFZޭK9ӻa!߹dA$:/hkuN X; ^փ0N. sw pE@ыn:'u+lS7~ uB{`[R+LI H\=B03 XϨo;ZuJ{={ bo_}gd)Xn}Gm"JELH2˞GRe,uDeϟ,2YedQMj=' 1pA0AzRAkʼn "]@ $p`O! ]祝l c@[K (aF82RKI8)Bj3H9Uk Kŝ\!2I"<ÞpT0\IƃfįEPZ%0qbh)1L+ D4"Jx;DӈĖEn0ZQρP)"L 16(g 5HC[ keYDB-u]Mk EdL1DFO{UV3¯[.2BI|Ȃ/u^UrBN^RqPUP5EnR+Dˠ<+c 94 ,AF-ŕ@$.ĔrJ:y]@nk*pA41I]¶`qCfc7_uBEb2Ć=I9?  kRaĔ t:z-^dMO i8yŅ pW ݲ{♮} XH~>Ƚ g4c!lO6SYs=Ɖg%P\Y>TK'۬c\=S?LZN}cdN F4Tq*A6FZ9$N $#$PÂY 2p6pc0fT|4\ĝ^TYz92+V2"xQey`&zRXVvEY!Ⱖ^ipyhV<67/e5cTj A| oE(nEi7;&; qzb@ID_î#A}_ ċzt?KFW>W4_ ‘20 wm{`Pf *޹A-Mag}w,"&0Y@+U2$Zi9=a;x^V3#cO4>޼&)'<*s.RԷW&C֟hS/5y &Azգ}'u{tAC_b2Nx)ɏg륪wU:GsA$(+40"h S\ AzwcKNc_f@fZbUrZjig`"!39?{䶭J/~q<<>.;Γ\ J4]I/@jFDI@PTFI*k }@wnJ$B$DHk 4g Mаlw[n?F2Bo.K2+-Ca|(ErJeωy9^{A\,]?up][\@, 篚|!ml/КTyUqA}xiVk%}j7_{;Z|}^W5Z!QRp"1wIzi[>Ms؜5avW޺uG*2nJf\!czW/^;e_G@ AQA jyC1w ˽jgo{ ؒ`6 jQkoFSPkƱ[zu9N12xKߚ/)\5|,<-.RBIs#d=nBn̳҅S@yy,$SaCamk2A1Te$:g-<:<ƭY*E#'))!tl]8Tpr֦u/o-tB |ϑU9+q$z ۝?"țnM}Dr)7 X~.].64p䶥 hznu&*8,|j1*.5 g\ DxĴӖQQ`isQ9(9XV)EG|*J]F1U}׏(RxZ;|dasԄGˎ0)Y )TJxeA@Θ>%;+UbPX0&YL!QDb .P)J AfuF߄F'IjPÅ|v_:>IDB&DլN 51w J8WwᶂVHC Ć1*"u3KbѾ&[L 9%0("Hͬ!5PO:DMx+CPBi(@'`T'+WkJ:mKUS6u/}9^~k  ]5^6&1:| *onCY^c;;~td,ؙvדC~ _5qOtWSec?L,șټ;{$ǭ%&ǚVTqR;@PjͷA.q_G|2YF>2g3{'l>{YlٓФ|Z!`LQP\'5bh%;q*Uە)-d9. x&4Yτ$wJVSGT %Tӻ': D寄oi.Őt&P=Q Sł p_0hF.İB]UV:2wj@̇hȘ"kT7 f*YɊ좩nnH`vj')^Bjj~i(_ ]yM~9,Sś-/qt?]IgO4QxPH,8=aRA{L !`=D!y)-$Z CjU\:dťZ1B[$rtJ"qH-"y!~~کno ,KZ<ҩWd!x 1;X^Œ~zZQtgVWT{{;ly|gu>%O7,qA)8Bljntd02 5#>YDd%{ BF-F!v"MvnrOߊ`JJXYuK Nwtdq2+8ݾ<1s)i8MWJ~1R^f1>?C1In7d%<ł|!R]bQDLj>[S|=0!e!36U|'("KDn:<@路~uZ"oBHaR'0T.nuHkGvS1s,WxQgsqdJ5D4UB(. !QI'BqH,P''Hj03{%dHŠ w9D{VeCU*mbDSCصU‘$JqAB PqhBK$~aͫOg ϘD@E *dTJ &)ϘZ#<5KRv fVb`WiIP&I8:H%\`>_o l9" jqaWbb9AP*j:r>^UNZ sB3QAUof~ڃ?Mgv1%:؛ =Iæ`;Έ*in]1ZCAҵ$pғ"?'j|Q!nQ  ':7Q+D[-Z(T-'m9!/$AF/$ pHCQR/.Vl&T9H_vրBڗ-ԙP%"M\6!(]W@꽝/.()FndP/r=#!V^0/ƛX<gI쐥 T P p5YHwG@~&pNEE^I\CTcvc{U۴\ic A!KHьeSMKTXQ@(_øC yӸ?R:_A0*[Tϯj-v4MT/c%\+7BB,)js~֐;@jBR՘L5L'.'8zLF:%-PELF-ȻUtׇ 파F:[ Ld' -]Rޗ UvژR)_ԄaejLDd -XbŐ3tx[dS$>]*#H gJ/P΄6WPlFy<Ƞ,2#d r%*R1p;^A߽yx<6 ubOQKUb3)UϤ:>wIyċ[K}FNc@ ҽg ъ+\ݠSԥ7uC{<Q;ͺ{^'w˳ٓs#@ZJlNUx&myNf3Zs(SFjg1J6Gƈw,c:?#љ&*(l,v!G,s j\\9{R.>׷D DsbҝLD,r>%,Ĩ@Qs{vA^]~ w#a#hgQu #,Z3ݩٿ7$#!Q"WߕF} J>|y``xJ؎Dq~{ѳv</Ru_[ڟSCu[//ARjT -CB'F`CD5$a)ZrЄΘXu4dH [d+[Y#t%-TAVў"VEF-p$1@[$af嘟%M㥟=9yz&0?Kϸab6p$qN18'Kk'IyA%v̉vZVsMɸG/EÖUq~eIR bpY@>ڱbHF8lA']Vx@Ӗ_YaD9WuYwM>trj/}Y+4LC ڋz!*mA"Uِ<OQ8PDOtHTS\Ѷ5BY5v/{g?( {f8̈́v#_b~xӿ~P0^tw:/abR3rbx2BQ8&!”X(fw;ܵ9%!M)p #2 X' P@Ph(0H@ a['j@&,\St&&ӏpQRW_A 66Qnlb͈bhfH*x,!J%)MP ,lBibtAg3{ QWJmaq6z8:'q0%t" i-6$:ENj tΚP6"=Ϙ8`aׁrľ p: 9mn`a.'X7ִ!0%3J "(J+04$NVN%vDsĸ!&QTI]$'nx9@S"R@q)0ѣO27Yא NT 95#X5TA*5`:w"aB8W)[,?c|MHoGrpٺgq38 7( ,_W̲Gt#ꗡ޵ 2އ_^w2͗D]F盾-dj~gbXPgo:X1 +{5xl2]v8qbz*kMƃ?gr[ʱ{˾|񑹍yxe2}yfSEnzGSۛ,s_罏/kpKqŪ Q\[dn̻L nb݄ !n)&۬{_M>jVK=Sg4d2pH ZP!0'GnO^ #Ē~q1qƣe\^|k.1-t1;NDZ(3O7IsZu]07|z:LY̌+'ʙW$1֘:[ t/̭Xd|.!))s+jd?ʎT>{5}p4D se=YGwpza=--F#۬7 n i#+@Ri>~9敐<7 W@!Jn-uR8ieHg[թ@Z:& hr-St *Avv0d zu> 1|C30LV7zSa8GPJH-M W8HCqѷ%y1/^|xQJ$IN`~p'7wɘfKEqi,#J ,v a*[/+*An&ej7hH ljRNM%G#(ERfgZ{#_i̗`>/Ix&Aưb(H5Y{duK~ziNgFs.=9B oymۛ.WY%@c0wB9#Q̄QB_jX&$%!-Kͼt%xDqḎ/ ^I/7_I\L_$/l&[/ 1I)#HZeψ-CxV$ExEqp2^kM48Tk gB 3UayR"]3x&Y 5E*ODuz,@+&G}G\!Y/ˆQ+LxTWѼh@dZ  QI0]~JIswͧ1WvGߪ3``mD{(̙"@nO$_1i) R$A1n@x|h7`p"}i͂{ϳ-L쏲{m=*|Xsvx5qbX̾ʨ??79[oW?ٞQLoFE9|=~OOm'Jxڎ)xlp<Ħɶ8a\2O#.:r䎸#zAMIxp6$‰ :nJM-nU|e7zOYpuߣWp^O?^BPt `+u'}aىzHbx+٩C ԏXnEddBA# :I_.t `RՆ9få)3Tv-C*c R@ϱnkh]> vW0qI ( %s(-Qt ]0b sXjQd$)3耙[߾?L %Tqf*\aP?y]ɯF²b͹b}W.Ь G*`by탇?΍[\{OϏh.=p2ܣߣټ_-Fy? tH9c,HXQrKVgg?|Wab['CCW}^ZUߜ>UnL@0:oZ-Pm޻PAHmaYyiB'QWo욦߮/}.|켡𙾟1 wܳo- //nݲ+.W <4,,?Bwώom/̙𮽧|@?UʵXYG]4˧jmS vG7Z+3薋A~#et1`[vCcG.<䕻hç"`"lno%4ڏXP>@Zk("T+8A>nLpCȶ[hQsw٩lȁN -ΓjjGPz-i[.jh7@x帄: BP*bTZ%0eDV( UqC[J:=%N$x^;ή+C+d@ة=_`ON aKp,aSN*aw㾗 .H氾ȞggsXpgj\5 `/zML$V&&$LDGQD."*s֝ aO,~6άeV \Hjo@L"86{ j4@j%|KJݵ@/0j\ŵ^0jc=S/Hٜ2F&P@\P;e ż@.Lj S :f |5SsF|Х`s)m0im(CEBsQ#a4BiQTHKY8w$6.1WO7j\`].r,0Z_B:Aϵfά5s餝[ aGA.˲R9V8 1:NS R y $G)aI@ ҃Xw}9NZ'I;N,p IDw;8V*TZYm3w, `{C5=da{oZ,lKq /D|b,&A()@3U ˤ k TQL4u$ՋmknŇyTpxnkxZ/ ww}'?['cfx;C-#=\}r{/K7[٧QN,V'o1s^.>P{ ?;s7I;4 TdQK~ jc B2_٠a㙌%cr™=Xnwd4 PnNFas鍼{qzϟj7nuk‘!o6Եs!~^{U_]77+sg/7OJxC퀄}ؾ1۰8 d-i]m{MS{A0@^K_ k+YO/RpRR N9 $j3x-[Xr.{Z?`X̭?+]SǼ8vO|ձ嫛S#.ձO(落j>~u1#N1^q*;JRuy=ybc,ږs: >Y a5 2u )1爦cBSt4GRJ](f @AK:wpj1Ɯ D)k%Q(APVNkюk qs5. 8.dW07W*b rM^E \=\X#s.\NJ&k@})"9?n{B)հ},.dq&0Y\'z [$ܔuJq%FHrNZbZ[Q')GB1[H@ao=(aߺ-?VPԱybgIEӯnUgl~L(;%?H;,;-fܝI<2)"Q% D\IC-*a2BH<UFw}p%y WQNE!&r?RɄ;9[fHԂ`9%\6ߧ|fuJ% X)TwF{cH1F)(j;5JI(٭| !0] NM*2 rnK $\ӂ YlV̵s# 5*qoO܈e^X A9* ^)'+ !:4P 2N6"k8~Ij]d#YS_|sF< mИ4֍j挟轗{/^ャ{o=թR9 ĬWF:00KGJ!m4Tg$DKƚ4܏vmAF&X&q)1DyD8T_#J4H$h#JNƂ0CQAplqHO`Hd!{K$ B1Q/'U)A96K̪ ~٨(jfA ү**+@T`A=H53>LXD!0!R5uwd2zȍ H sU #fT3Tv0J uyZ$%VP#Jojd ,/3_oVQj+.&gίv VmeD54nTz݅G Ycy|.hN7>_p$ X}>ߙ.mּ"XEYGڴ{U կ8;.[F*8a")(DLб8j |9A&Tɢggkhÿ#7ynjΖwG_OjѴzEf=̦[ݑ?οa#~rqk^7cs!6օY4 ;ä6Ga5PIꪛr]EOw@#$jo@*2&d09SoB$V$4K,:F> 0e(P#gZ& ltT|Nx Z: EgxL_ZN}ƿ se2 6:۴9;w_ sڰsg`x 5Y^lO춏G+w M{(rg{U-3ݵXM-|SSDv uMR3i"mJ޹VW1 R0*din3M6PO퓞h3]npvQh=R!0gC%`s'X#۔%sتt8Nc 8ݘlO*QgVAGs79^LQ6%OZD# _&kZ4yw;S і= A;!^J~άm$7gKKr ͋k/Ǫ|*boҮ5f2w_cAJҼ$A/V*o;:QvcH>xo0m*A 1+w@|u_ib׼`ZI+XZTJ!/0nVCfGW}m'Qwxvm*ƨg>;QUʽC:l}z8Z-oZ'T'-YIEp̄||b-6A'm?O)R %,(M:(mH*%'(Y8,oVX/cEKEd,>?R\qEA$Ԣ$#Y&^%O*ʖ:-Qw+nښ/P[ D:' L?O( A9y4 [6r٣"' ۫RCn-f!ޕe>XQ?w"QJ!`!BƜKDiAD ^@Ƥh]4D% {($S\"|$sVd6xk33jg#&1|L"IV Y٣(V rg6@a#x7Rt :9jAT6!D!*2IJJ(‹ekڋ&CR:-b1 `BdAK w#Lfךny4EAAS$/EF6l, lAXZ7+wc Y>l >F6]昈l!{,wSLGkfVRiW*sxky HW=>V^#{7\Qnʽ'r{n}ts!6.>}3̼5}utA$c4j =}mD\O6@=q.IO[;ݝ.;vZIrcC!pC% |PpV#afrҷ&%mz.ʋ=*VLh0B9̛hrQ9<;OM|/_ gP)g+V59`4ZM7e[nfHGpfs{18n+Pڵt:r9^}F6hn&z [//i2hsJ&Dvu+!xv0Bj =+f!;פ'P]:7B#^@iRw쌓d>)N] YtT3CQʵog(a)/U8JN-h˝4Bk+3w"4ۚ]/ovW]BV>O_j;H4ʣI<1EI껥e/GpD.ٯilb a\/chZ@aX:18nBNHzB&6:iy.q 2B2{WrϣZ\f˛5hXtS?^y򪾴vruvݗ]W׹>z'}>û>û~p8; sFRBh%8@ًK1B/ycQD4':Gy-~{}v6ܤKsǾ#5-a\:qXow]o=M_kuNח4ʗ[Z̖}НD%CC߲p -Yy%QNEX:":.:ৣgJip6!F{j:BxokP*WN lh j`56('U3І0#ԃ'&,Y!A}9һPf.ޕ#ZhI'BMmE|}.f_MGfy%bs^ eHi1 b<gNëW;'y"yqĎFy;Y[/f#O0Id&fދu VY0y'J8gHB>|`|Nˤ|>M>*_8-OX<62w\?DfvSSA <ך@ƪ%( 5)690b@NMa RuH,VtQg|x[0stvr{4441bӁu]k]4cȗ''S=^2e PXtcbD`-jL..8AvΕnTqt/z1[-uIilJxG%yg2Q3, $dSp)6VNJ%i 2pacV)utM&"CI$3er(B&FPBІ{da ۸(*]Nx?M2`Rk2 3#v5ތj4 ;d2zhĆаc\$)Q[TT#(b3CŒD% TK4Tlx8Ȋ:pCSնՙm$[2^gfA2!Aki[,ڼMyb\ìvSa&'{'}있2c a6+': $)O&OĦ v)ءyKCpR,}tZtnc^j֍V3̮;e*xAWz >Oa yԧ& bbk4һȮKә 6+;Y<+v'!-OX40o3_x_^l-F8-Oyx(-4 ىCiaKpru)]Z:П3&i?/y`I{S3#N})Լۙ9o~GS(n6^3 >8;.[(Ș*JNӲA?4Nn-Sجc$3{DǿMbw#M-ْY-pE1Jg֮ ԛڧ#~ƾ_sk_^b)w[HccspЮ˺$Wmm(U$?-zj"t~wt:-b x޽qge?ϗ?k3zo;kcE)GGqukwIjC=i,bouN=p_o15Elk\KlM| l3B sr^0e"ޭũȴNSD/A&Tɢ'm|1KYm+fɃ~5Zi50앦+FE<{A@)iQ̚| LX~N? Q$:&lRفP3 <bƨ5$(N۰Gbr;psSdϮ/g14#8=a}Z܂nS{SE3RLoQN=07QljyW>ou?}4VBmJvF`pc&w!qdqUק)]:OQ)gRtS6] 2]ﹷ`Crhy}$+Y|rܶ#s4)h; e)+kv&zu/_A&%?kӀbG(|5P3f[O5XJi`Z0SCɓ뷸nv? ЈhaBN,Љʬwjg y;N+#$-6uPҹLD_EK!C$<"uw"rq% t%/La" tAcM{8}0#BK`#&$$rYH_"xW@$CZkQ/GeY@ |DKԟu-_-se:15dUsE(| k:16e#+8 ;QqH/U Oy7y=At1C<asTgѐOδqHAUyUVM4eL%xrm*`1..%#b7xb07B49 L݄]xDB Qx!6O`7U\We7V0n90};&^ ip vTHbtIۍYk.<>Yy# {w7"csɴҘYYsں`T Q4ʵ6j~0RN: V>H{*Bm K4@7O dh#cYD $֕= )wh%X<ӬzO7_}osn=]FZ#,hر}Gݴ h o~i0x0c އ]qz>j} o}5w,mɽ:nݪ.Qmnܧ\{!h TmC\My7ϙT5k$B+}dNhcylukqjFvK;(P9h=p4+va;@:Q|]MQ +>W9T+^yLRy&[џj)w!W11P5$jNUO4 S  8 KˮTkM1H̾}hto/gY<pncV{cw\/&v*|.l|_el7?~7{H_>cEydmv -:}_%07&/?i7l~;sˆ*Wށ,+76U8:Ŋ <+}F6N-5PnUm M4˦>sۻ)͎ trŻpf0wo2MnCXWnm E _IJCK2O|a,9dY8cWfY*p$=đpZ|SXE(Z"|HzJ.b{.Wm&9%^aŁwrXY8{sYXwsq2ep77/_6$8[ƛlj\#zbcl/NuljKKzj/zg"ܣQvv-R'#|X>2}_*iR1,{Ih?nx'LX1$+^2SVd UQ Dp2F8J}o&Ws<֮PG/ת+&㣧֓6Vt,&%_6=ؾ0G%ޤhъM^X(%Җa;92a!VRe*ϼϔa;:!SHlv^Hmx8_.9H-LOI2|2H. `)UF$#5ИZ5'%Ks 6~4!H~@c%å|fٷ?Z,Ϯ.1$WfnR1vBgaJQ^~.Oq\g*(ݡF}N?عC+J>"(%=8<7n`L(laԹqkE&>z㻪5.㮯x7ovkk(ʟDْP5A\\({_&Eƹb]OY"Q\yyO 'G# 5=ۚ &\Íw쓸J?B2OyW&[(J+e).Zq/F+sAɸdw@%'9)-Uj[P5e'vG"$\i"%=j%kQmJGbI4*Ɂ缝^ԥT#Q6NK!U)aO=Ch3WxL40\iJO+SRl%靧ԮK@J7GAQYkA]iAꘖZT"5RkƴM&Jw=)!)918 7-oA-ϳMG߉>W58}!L`ly (8: E,`@}ߡ A_  Η/,KRhE/5v&z)#jcp_G >* u6wAjQ$~Ь/O(F<},KFG fuڅlJdGٍvH`Z Z 1A>aKݓٯ;[zYD4^0CgJqtgp93ӺܠX݀4yIMA|-fL3Mס=ΑJr:Q#-=yϫJqIOhZT͡Bǂvy(9c/U t֓I{jNU#nKoϲ)f?IrvwY$y˻XC۲~l5ZwoG[soi"2Ey}\,ó `g=H d ɘH(-ༀhUXBu Bރp& m8}7 eJMkU rN$JkjlJnj2Q0 L,U8ZpUI)`UTP 0'ű Dq°8#^ أ'Q 4:v$9"8h;'4=Q^eGzeDք"anS%=pcPC䯷+u !'z5wXPն23U) .VKC'6--& a2۰T=YBD_?߽Y7uݛσ uHڶX֔⑋ɉ-NWu  ͉̈٘*S-EX}[TB@5!%[uD~϶C8X_TZj$ U\]]Dz m~ɖ/on&~n_Gh-BHXt*~؏:JX<<-{ӿW:V:ȏ6geC$@hM)~+wJw˕A>w;`%Es->ѻ a!_fã8lreb:('nM%vg!Ɖc\У8 r+Q&c!u{n׻f,%H+XNC[r: ÎO=n:,soʟ Ji ث[PRnjiJ-YJQJvDySkDT B3]zw$Ȅ[ $*: oX(ܭw 'DZ_(-Ұ k7-FMnNڒr)GIVZvL\ 򌗭(A%N>&I'_D'q`Wꦁfd5NhޱN^(Q*v0F;1ND.(A#Ri# *A꒐8*~GP2m騷j]y,79A[3D 8-zZ؟$}٢+Gї'-0-9TxuXFMpBfB:-`b:QzZpςX LjR _*Q{ _LYDE ֍j1Ղ)= As(-QT P@#$!A˙Zz@A~T+9D` 7Ra B]*b *PqBϠZq.-D`g^BR{&l7#_!b#'X,0eB%qL4̿Odu]YYM-bV\WFFDFfeQ'x%BP@kE%r(Z)m[*q6eδ =8e<8aDBMQSuN1 傢zڇ@Zᔥ(҈V%HuX$PkH' 4j]+Zt?QV8ے3¦vh2%x5IvNiP "5!D i E@1Gҡ2IHb X_@k5j]$$Fh-eQJXt-okqʰ(b GZ 4(\aƣJv (>VZЂZl`->I -BkDB:2~\ևMImQ;'f( qZ ]&Am%Z6&4GHcTFCߤjj[ܤ+@ivn+`(CÂ7$vp ZlVkc!FD`Q$@ /QAڳYך maK@*pmQ!B>JaG.pK'?/$qDɷ8VHV܂0Ņ%Զ7fqH; @VН_p5+c֧H^r"i1 hJ.Q#5լq{m-Ŝv.D^B+#$ninMT_&44V2jиkc@>gr)<$}G0"<( :/-0'7;@Fi2i)39W|PpC{ Lr$\}Mey# htXbi5 ZϜFqZBCF[ݶT-$PA9x>/.]jYHm0x/Vv>j?.z[3+ Lm'iUB]d#E1\M3*3E/il~lC0KrNyLU,tM{u=ߧj+Ydkr| Gx{C\ |j'h`N*m3-yu+\JHΔ:c+N '"M)50o͞gV50nk>^igf_>=;wQ5 *"$ه~M썷ǚ!g6'%U?/FyC9QJ@P߀ZS33HJcF #>H@ϝuP*{ :'O(!1CwߟX%؂P+zlI-PszUM#v-ANKw :{Ô[hц'6CiAɱءه9{& (O|-FWVR lz`}s)E0SCe/tS-KFl 5o޾m\_!EghIbR^.9t*aSVt w J\ܹ_%`+bFNls oC@].k4*of\6s&tNw]i 0N0N5+H1֦6{j{gV9@l:`ɺ,f~S\sΏ‰o ;"uSOuY'`r!>Ϥ X>`2^MQXAJ1HW9!SMjV)gu&RzA$Ĝ\j& !^mHWt8 XԊ }*]JR).ܝ{mSnB J8!|?qRfH$4<ՆfBet#L*e㈠ Ci{/^LX6M,"Qxi?QHㅰwԈfY,D9Ej Ozj1D͔zUayѿ\3 ϤT*ٶW0'Z1ٛzBAzf~jf^*Sc/D"tiHl#.u:*Jj[%8aEmg WZL4һ҅F-/1,잋:E2碮00#B5Y94w.jdR H{;ua<6|k=_梮Ħ#E1`nvx /bz R2/KrtԿާ/IK#g1?%,׏GlE?~5D! чd> jws~մ12Rj?zhX0s1bI$ )R'l~XrF,xfޝnl}fv)lv ;/,,S# f:s+r\}SqUJs!z¦> [!՜ H hXk ӌ]ؿ[,=us.Vϒ[O^"#$#MKS~&6ΝqTY`!HF "Qea$vc157(U$c sEq>ȬCP߄$wuds5@J9cY a_b, ]*ҊQ($Ø;$x˙hw܅К 6;GjTQ B3E4hwGϴCa=<E$u2 1TɍT)aG U*?ѐ wdLЎ$hnhx%yw."/6ac)X{1ޜu_#07߬KJn4jϬAڋO̍Cj:_AfB@U%UGexp0JWK^^}i o7: K@*mt6_U ԂmjI585NӖs0ʖR;gk>#5ǯ{w< uxm7Do߳ g# Of%)C`6zz4}[3\E \1[oBKH` ۴ QynOiAF pD*p&cYa܌a]yTsz!n 3ދFFƅV^)Аj ᕑ yE/k>f[lҿYS6QRw#̜N>ӂիk\sn2.R&*XW4نUIt|z$3z cUQ'V· XИ40[@Yq+YTN qx=f좼JclpT ԉ) M c^3׹չ $AKs#&T>FC_B ba4^gPS1GC kA]\b6j|>/*)U%4wUm(2P .z j551Y#ˇ͞ +jN^_߄Qs[Y""HM}uL}|nU͵[I..k7Jˇ[j|=vQLyyB?}P}`*vBtlګD;Ė3L_yӿzK>zIUgnV: w=V.k*I7.E2%u zFڭ9S!툾:̧djhL vɶvSRv+A蔾cv;b̉<5j6V~6SՐo\D7dgOuIjHuH?niI1+  ID= hכ&*L)4-] E) J VJ 7 p)9\(¾B!+EdmNe"B 5WK%Ch[]kҤMF>gNHe@9E ML̋%7%. 笰yȼ7 Lt[P+jeэHnIV-%_ #'/i;Ef'x L]Vo&t:KHKfq]y3* 9'8wH{Jdk* Uq.xŐ}a=XUg={a FC8*l^mWb|")e˹C3;/33w) \TTzOJ%(OFdHמ:<]I e}}vF]'Qn7*ؕ!'_d/ }-̺%:ͧۻm7٭l& t ֫`vz3ڋ*G*7+ `0)Uńe [P~?󏗟{GBaKzSŒ{{pٷ%J%Q 'Òuc"2l1UC%FC L8)J#;t|EOs_lrBzu"؜<$a|||xR\wP]ߧF/_⇎?*+f o_5KSec t}^b4yy{p%_ߜI%)Jn&}c4?6w_d غo= y oC)Q0j4 @k1 ,^b}M9imTd;ˢm*͛ZP %3p3oן3lY :n|{6v{KzwXKIN-ock y\T$PKToM aj{㸑_^p3b,`? $wI-lpș4ycO7{8R'حb=bJ%QM1hQ5 c+ Z uý5eۧPxfdÅi\5*%ԃb7`e! 6`Q|af:+-XA!'0uIJOO!%jځu0@jk|FeNCcskwK*EPK8ИclJ(l=bSqFZ72,Xxn|yi%uej+70錣F6MS[ k0;k4k/>X դmՐUteLe:L=0TXS71 aΔ(TFUISw: !-Wu*{Gk*'3@+To'yKuQ./ tհ&NP#&rÊ-s)(lA˴0U D{0cffhVs `klа#7+Y{Aj#um1؇ 3̑ RZYUu-HzFEBL*X#RAp] \8i3i)(nٜ@9+RZ;ѯI R42ٗ0qw0۪H}j2tr Oal!jءJU7vK}&XjW2V84,+|Ejo$#SFf9 < ui9PpQnWVc8xy*B+5sp%&3O/hia6p*s}.F/vPg?w=<38tpnTZá ]@\_# kCPn%YCZP;} 'rTجKNin '#1<+3N0'כTbݍ|}5wckr>-t= ݙYؐcGcqå< \'Xp &a[JQM6m)Pf;C/Xi9WWSؽ< +qSYńAM%+6 S"j~SͶ*\:LƄ|)ܜu5 Ǥ,Z$X|}T٨yFM1aI?zDcH "'Fao.8:AR$:5%odȩ0tD.̰n-5^}:Qz!C/V4G䒜`Lؕd$ֵbf&5Sřai4lFS) YȌ J! kZJVMJƖ.gz֧|hJ5W*9s-i hT0lZR8m>4mNol1 #i LW>¢GZ"3p 4o))$Vjċ(ytX7|_z$0"E "L]382񠕼2J'xj&ڈ- wq\蔲njlUO曉fNN1UJbk96 xc*s-GZmOw7jNI8n^|iA\bYm l .Lsa aTsPxlbE]1β!<q)k]"Ĕ Qr>"! ,x%]z ~0ŎSJ(/k^ ?RF<'Pi@ ow baMBx*.i=6r8(¿yg~~}&b_'լy o#'[\3~9 s=/~rMp_n~3{tGVW2x 9 ]=q5`fPQxzOe}yyp0E ՚ג1lU!W6ZKfaZFIEx_YH[n%PbJ s7x 8BȳG3Y湤xb XcɈH[2TOxV@cܯߧ;$INeq ڊ 'Qb,e/U'x" 0LfX- 6а_֓ؗdBkJovZ2҅UªnԈU'FTU8۽naۺAرFJ,./֤VKug>;g'86{0&%v9Jal xٛ}Gw<@ٛڌhwW#G`iXϼC9dCԫ9ݙ{`7ى;d"*-XN%2ѰUCΌ.Mk.0* h!/MN18h֎R9G 2j"cM6-ś }Nbyt]KnlsԞ茗WW?I_;O4ײN-O\6ukaƲ(RFxgbccE3èrqd`;y vaz^?ú D!$Vrf(߸7*O׃WE)Z u;QF.$C_/xgkISM갖_t9Jۉ&RQY7!Iܷ t"\㒀B><2;i1 ;qN?XM=&e ~Tq۾NU;ѷ;Ń@4 `Exơo=^éB@IxnȨ|__ϼ!Sݤ߻Ƀ{X՛"7e&a`Mz c~@ B I|%F2-RR[`6Y,Քkƹ >&q8RhgV-V_)UY[LIiXٖ./^K i;ĥ^) EX./^KjVzV̀M@D9)Yq~T y#jFL<̛| wj}Y~W您KKD7mgF1o4;0Yr"je Fl(x΂ח*%X,C}HBNAjxȂO1:?CƮ9̨-EKPQss zr6r+Sk_]I89ξFBDSuVI$h.hld@ʛ`=gQ0n8!rDE퉟ؗ'/xB~h"gq/kݼl9-G>=]6nvt* )'W<鬩R8 Y$YVzvV"dQ`Y(Ȭ漊U;DݵSdpV#%hެz2}S9''U]3ZjwfK gӑH͓3m0Q)<7Js1.a>2|gNrdb\wiY4,~p3grQRUΣ48 TY'h) `K-&ϟ>+OuRd?~{]kͅe}>KbJ`~ڌEAvMK=:k϶вmmCt- P lb6 / VSj3Q;Ѻ̡[-||c߈VK@ G:KsGÁ:2< -xP¾Rk}B}JB3 cӶ73In5V7}q¥! :8!#luzmdDȄܸȈ n:δ|b}2A3w 1Pbc .bư=7$^?eWӻu4R*vTvr&kJbe>"M$9_)kr;PUkj"@icrjRM;K@0eEżޣ\MރULЀ`Wd IE!}[*xSgJ(1>l0jE37$HYLjT$R>6dQ5'#m#Zk-ML*Zu}^L^c⍠Ph*K y쨻KWR#ǎQW'Zug>;.r?s]~ CD5Jp-Yb%xr"Z[ sHc,ݝ}>;c1ބz]Ks6+}qV5 $>l5s=8Ʊ::H.Z#X%UPHD`"3KD˹brnckVsbû 5>o0o9"^s A$ZVXVh]45!W]cYp U۱rn)HtUNv!cARPj[ ڔaeRKE*fWbYHR3fR+Fo JЂˢpR7h +mxblLema`JwFOm3OyqwiQEQ_QM\;"{1wֹ3y%z)~^ʴT&^zLVuKR9ra\n3KӦk˺u֓g&J:O^j{fK[U1Duѹq{F !m,*N|@FRRM0}1k6fhq&?YTp#bhr#tNUS^ 8߁y ŸZ!}yň QI"-d8@5 SQ&wڋCtc Oc>&2#iX`8b6IsN2-^(AsmĘ`DuǓ8Z7Ij%Ϲ% Dۚgv,6Yr#%+ǭwꐤ<]/@dHgtBGf0/4:*iq!^HI =NU UD\.xM=`$[hs' /[?`5͂Y6Xn!G= 9IbP;f5Lvn#5wcrpdW {=לmvip 梒 'W-MW>:H(GwGYQp6 I%қˌ/hhB{ߘ?oi吋|^Com/W{*{@L< lyiݚq96Z'mvd=t>Sj#M=2M" Jh$_6\d]Hwaz_pD3~>iOR}k8PGW_ _ F yT1BnfzGjAN)j_>缔rKOH}#3D}^JRt+DV`gnϸf?ik +wo~G Ӱ@\P\< SOSBɘ0Ѩ_.<]Ilw% Iw:Ѯt|Xh9g992K_w؀%g=yOo60A}yb` EԆr X Ap}~wp}.:'QOѼGF97v6<fF;0x8+ 3hC."mI-'])?@nwSgZ[ ϜOR)|LRhRyQRZd_A8c$h DփըHJ[>I/_=!(_fY[y/n O~BUځ#kVs0녖ql_ۗ`:C΁:XnNr [g^'J`Lf`yeb`8wE)¾#;2ZGPۛ[uSUҚ #0Qy敩 B 7փ9gޛRj8HU;;걯d%a"_Q/ 8G,\8|~BQ$g/@j_;U2!h\ ryUJi?d49m.QL8*6$JwRM8 H4*=|2{TOL9qCΛ{PD(s:;"OkP*ϳ&eLb #cd @H8,iuA,^{Ø%~ER,JZII^̒R"64jԋWSP:E:,HO+'k}D tÖV;1 +uR .rpx 1w&R  (zWj4^+xފ ;R ӊ^JRJ¤1;R B)I?/eKeRk9|J>xѰQ@*4khYјPֵ4(T r;ïײ[8);>^٭g:)fM /5K}=nډzړ::.)/;z Jj~j@s+EN-9UH13{*moW8G81ǩ ʊVr,?ˣy _źC?b Z"1{SbDCT^ 3H-&i(Nbc,pr*}zp I(1O؂1+9iZ20tinBH% %3&=t4{Hj#gϧ% kcdYI}Z`{vjIw{cvrTHmDԢӨ/7m4 {!v%mNKL9k*S}"]졌fwʎ}uWl?zɩKauu1#&YpMr"S:*oE2G6 u{oS1$?B1seCUӹ*fH.tM.[=:>">z`q?&Д FzEs7ع3ܳliz3oU%vCtܗ[_x @E/ v{ Y_5b&T2̄Vơdз':/ҿB x `b_^o^t);BԘxZ=T1iH ɾ)?.rX( .~S8tİޔtsJƅPN$Mn/K4R*ZSԘ 2;AȦ,Wr@7 A =8%,<'KS8h`k״Y :eS4tP"q\&"bZs։9ri= ;\c.z~nLWcUYٖӯnn|^Yeawg)).waVh$UR(h-Dm 4"PЈn^=a!Z2{6M J^ܚBQ757€}JUcO9ҔݲB-UT24=*)"P VY*q?Eۍ~VYFi}r:>kסۯYRZ|fB1+ҀnYO~ҚwK9 )e\)Xgdh'rwn%ZޮjǮ?F|{S[+L$v̻n<ᮞ}}XƇwW6xZ}zN bW z.nVլleͶhYom?uMƇ{G_ގf^~-Kf8]tC޺kߺûG@`*7o[}oWu̻!bW Â>ع<9|lO,֓ >9|YvL?K޹~ݩ`[;V(bx⻇۵DJHnt`_.~*Qzڬa~Ζ:՟)s[v#ֿ=Mwaij#fko˛4tH:=sٞ^ݮ3e>E)f-ڳCs$l,ѪC,ЧQVFQ 4+h(j0RBem'{J]W*MPFRk[ K(ĊmHɜ]kI ocԮ+O kTR(Wjsۙzg1/a&@䗝ޞZ&1BhdJ ¸+H%@f^yԨi>rMk6xUY=Q+eztNY^ s:G6Xr495$3l/ r5\ػaM/,Ʀþ bDi(:P/>;z~pH ҿ[Riq@w!dXbw`t&8i;- 3\f8{'3Z_"\ݓ?_}** d*TEYy4nG8IlNӜyZitnRH{פAp "Mk%j Vڑ>JF򢱊0ZGCQ.0HhMc6R'} Bݸ"3.A1U-&ߊ0:4੬c|qhu|Osuhus5NeroV(PuS*+_^|R~+E~!ϸoK^@ɞWύǛ>5>n #9o;Ry7͎혹_~u\nuYtő ؃35.AMlP23zhd+CP)Yj[01jPy?I#` W~;x>r"t2(z|[o i>dGg"@D6Hoi'o5Xʷ{Ԁ=+wǿHJpF)DJ T4@AS3lO@0e9uǃ$xmڅ'R5rO)9ÎiL 4Sh "a䔲I vè˦9qK2-ȼ&V\⣑{nAٿ%d,an,q'wf5>iPI1ۥxap~P^?{BWz[gS:ZjX:ިaΆz0C߶~8_3,b7Qsa n1,t3yZzκus9ц\o`{λ_?r*~tʑ5Fl~Ŀ}'iT*J5& ʠlʚKkS ( a4\kɁaY4`vS}8̨ )m,(*U3DHS}(?x0i;5Hlx.6S .S "rriyvТu[hSeB`v[27?N1}ԩY΂XbngtyǮgo^`_爥*6 h+Cu]*+ڭ B*A/S2ickۊnACWu,Rk[ߜ>׻}UxtqTR^WK{W6ow?efeѠ$*rf3GJn[~'MI'@q sZᷱn@0yБ)?M~1"6'ٟ{~1AŽluݛAAG{쏥b\{PfVSP{RA8L"²8"g'H& Rǩr}pp\Vc)5ɬD횘ٰra^.F^ׇ176 w](/!G ;EfBǩD("Y#bfGTQ*D"1W`˫ξӋ0b5;\pG.@r*IBN> ߀}|nJۧ^"? 8jp wӻ9:ٛpw/ʗ۠NY=܍Z~{ۥ)6Ryu6n{ xx:NR:Y=ۣɆ"]J6dWֺO6l9Ad#H>qtVs5F3'[Y]*1x~41m#+~$`0bb14p1bDNR<yM00 d$r3C3*n- ȹt:˜l8㚓|f՗[ip~.cEڍ7~[܈9\?Q:a2Y9$BdD2b`(k%pr`LMجfx37,f<ˉ헝F֕mw骁wm ZR, ;.PȎL9mot`;B-Bv>E@2 .Aؐ+ %4봫XGq (! &(0٣pU\>{ՀI7Dc+S.t:҅q5}ߎVԵoMA'7Br}d+X?a*9'SZ)ѡM(Bko҉c2+(8Evhe0ʘ#!f06_23]Jj+x|8  E3g)>vy~(°VKJQ X@bLgTP$2I,$JƮ`0业a٘ S)H1d _Nr>6I~>F VOfA=a]/6g *t [)zll̕];1[ Eh%j] քM'_TVf[<ʭPr+;=|,է2)G|.[̺֎> }֙j6^+r^+xu.wZ1~ qxzVV}4 p\6Fn~{GELS)`-K51&)j\ D)ڍcN]O책@`Kb|}BщEJ2 _}I\>A/uU^#y,k@ke,Gu+ݛݥ6Ӧ՝j) /NqqQ^< ov{m^Pz* Q6>Vu:H3%e~"kqC￧ߊ^$t9Nj6. qCiX/YA;6u4^ө^{nwfU2!iPT N4={Y Y?)'[ .IBs ).;oik7p[(.v=rZ)_дv ꐐ\Dd[{ڍ- }v.[n[L9$CdKҢs̊VVzt>{*~ƫ{8SCE͇Q<\zhDSջ/<aaQxa"c@O_p7 '@QQZ8S*(}|WC v/6,Ea&u)l+]VVN+= mZş9=^[ q{Q+m(Y<܍=:6zc^wZ^ߥ+g jbFAVMnOp#80x,IΣ!|~PMH*j3Ea(᷀aɞ<12?N0q,.= Nty`\`գ ka.M[wq1d7{f+zRC1pqSǓz=j9yR-GOjhLMV+jP \D'>z֗rҴv ϼꐐ\DCd [{VDo N.Sޞ5Ʒ EtGL Ss;z>‘(8Vx:^Y2WagzGg{ҁq2[>>T֫cO>M&Reo?7B-Oz!S*e \INJ< zuJv:(dNjˇp,;61IAT*_Au]2T._[yDK%WMT _kԪ*WuZ+uATO_`J!tFj͵GbLJ0Lecf7-L/fK[t/p^'f^LcmS8ҩdU o˵ʲl>}_: ٹӯz4?٩~(Y.s\E unKd܋~Kz,ɳ`=䚅lG#Rb(,2 G1b&GGr} .-ل64ԈEk&QQ`Eܨشy{ [D[ٓ~6??hٷ ]OY:"*ÛORT~lax7:B='kթM"EyMX\'G_q`ТQs|>]miX@mi)zZ!P$sw6o 7 @7ww0 !|f!嶺ᆈ۩G1'5n#oI9-}1!璛PaK|ӛ܈C(ҶI> >~ DPzz$+.VxcS/#'ʦՇTq 37f11hjwwN@-*=:սO#qb}?9I򮺫][SI,16eq.VXqKṎ'0Yh%#]HV"O_^'qBUpKX2ąUq&Jd"I51^"?:0X93-q"QplU$b\̅h!5pe@:s"_ z[2$+D!~Y*CCiaҙ1U(%1ԌqU)W ^>(JܮScEXkAYϒy{0)lؤHIh63g/g{FWά%!ƚAى G3ѭ,ig_u(Y"%Q 6Y]yO^UNnSMX-F$i[4`K@9/tGw$xw7/BJ=^:S?=s B'e4ՠzY˳?[qr3=X30^i뵣h OSDڙC3궣4b~a{lzovprp d4;j_F{sNqw߂\q{[;26u;Y|;/9=[toKg yl}o *{w$͹keKg/$w} 'w{O;_9#p|d̙\ݕYL{n}h3g3x'nT,X6zi :;O2I\кx:Ҋ.ݸ;Z0 ` w˼y0J+ht<Ș ~te;yVB! >~<`(hQ03}`h02X C-"MS!~S#OqO`(չ8D)4,ܙJcR%KB5,Cn#RΝq;Pz(-HkRPfJz0@)¬(}i?r2R||x-PʨJ;-w7JC)ZK?T+tKόYoxOOPYPO(=jR"V⹤DcTwV]z(ZwR?^3?#G)~(e٥1;VcF)n#/Զs'Gu@TwVN7J1C)n%JZs?bN>(?.͕d8$JC)IR|:n2R.eQY|-J.${J]SkJKm;RZ'P͑8e:rbRtU@)~(TsD{:nzev jRJPR-)Q=Q[`lCs_\]==' e2$6Qz`zqSK K-U@'wT /<'kVX, &~12b DA3dldWϳW>F ;sfE'^"D\+ &Ao'OD|&y>Ee# OdG8Y2Dklx{ʪ/wCE@tuNsu!hTFN>-pvҥ'έ-)CJш))uĉP(Vp"c~D!fTGfC,24>"[ξŻ ϔ2_֪)ԟtqj`iVTdTW[p҆BF82R&,6Lb k.PETq?4lj W(Lr LlDJ5\>Ѓ;@V:-Wmas$nvҟ#5\F:`(fa ˜d0&izԐX$3u=s6KY+G$76%~ GZo?yeq D Js)wL~aRav Qa$8Ynډ!0<ϰևߖxfW\pg>{(Sm#khw"zrŜɅX scw=1r!k!b9!_ GyH38;3N3 q .tc;왍آA3vcۍUl7V[h4%1LC-(&JM"cF6 "(+4'mn{Vko,qV-e_n5'S;:Մ1qlL>K]Px7.et2ZSEXB? l(A'p*0$‘!"@g`˂ $DYY.gg_ݨ؂5:B"m,@ FXV cDERIj<c'd7HڠaŸ'w)˲vzw㰬2+Ca~Q" n}%[ AjKŖbŖ曹Bj%W1 &i,WЈjZ-ia-ܫR&I5e 5Z--ػTzpaOaܙHC7f5ЁB0Ph$i51E#yكz[kq1)<z:^x 2@=Wzxn @[*t z.=X#F3x(DahA1b$!`F 켄G42<2_`&n)$%6֔x(`ƞs-duvFXRIc_NU^mXANj^xP'emJWp|&ZTsbUiiїZ0jf8gv{V 85۶OݭfXm dN[=L0j?;{WgE MFFڰ `@gQ YAs'Em/mm;$d!kK]NêGkxP748MX/b\<`b{ k=@:\Zio\Ecİ4Ex5Sp1iq`wlzR\OF% ))QBo1ZW2BlRIX1H!bOAE< 3Aíoka!ڂ;c[as2NEY`S, q9qRt 6ba-K}{Blˤo38mO.šfWzb BtK<=Ք\غEWYDԔ}GSNB i-o0x}UgRs&1ney(gwY欃]k)B=[y-)^v%y ]EAK د!+""-%UaB}6D>ZȦ*^CDf(l vuc3gy\@`}F?jvYɘvuрix,$0#dؾDm%HaZlݠטDr,wGd:9H ,|F85Fo퓟ޥhe2v=>;)%!$,lxB$?91aRӘP@,4˸WŜ v*%MOD43" CxXVGv"WOe{14,c\[e$A\!aT4e QV3f+ƉNRi8d& !˜gR=2uleNT4KDFUjV6cR*K!iZ>t p(k_?Lc{KpC_6NϽN̈U葾[M>o7S㉤n_˗,jUbn~v6gc0D_3zتbțbF35)ԕzt} HL#qubd>1vPt&3 vMw77v*Wt> .h>~ÔZM'PtRK wOy 0#UW"۶U_Kz7 )g0sNJ,q@If'BXĂǠx"_1qZPT_A5]p)|9uٮ@nŁ)Z&VSR(A%`7}Vk#XKGNI `d.STU۞j0@+x"qEO]ld["ݴeT'*Hv][<߻\IwnLuPP3z3=P`aro}apWOo}鏉$H)x0Uz~93SlHmt!ꌴG"b'F[Y:^Z?59褹}}R]y="PZdع*ŷ|t\cb}OvŚ+Rpw C J^띋qRjfg+L `bB.9l$oQuU=Cg-LKQ [NڮICٰj0Cb PSX!H: eoW!>:(*,J{'k&3CM+) ܃ZA4Owi|mTZ0>-ې\Dd 0544wnG4ݕ} 3<[hL$C8d->:혧̮[xAOֆ"#SIILNX ,r=ǝ8]LMn6>?;zԌ O㳯lz_}\ ^7 gwo!/_ Aލ` nM~ 5ŻN@x Nhp~Fuh-S`sT8ʯ*~T)?H "c%cJ&Հ_2&<φ5HZ|(ЭX#="ɕ =aՒ^&gpO5%A@AE;r5w.Jd=1 썋A8i=׭N_]PŽ >}F׶k{q"@lIX:.>aPk B1K@MGv*x 7]51wP bB;/;*!|8UŐW Cݮ fYήnl挅ccכǿ?T,-5[bį([#7/F]ZiMrv<Yr9{qvXL /PlUx:Z!~c jn<㹊n(x| / 0moݮ{>D1<&!ٝD$xEZT2wޭHk`ߚtIm^PO^М$MYIx|ZB{ |M"3\? Qa$LS\>BP JC+}μ*?=>Xf+ǁ ,a\h!!Jh*6c{Ķk2I VHF9JƂd)bnE]~utKsj:uGm W5Bx&(:f Z4F&X8bd8BHDvmo;_rgK7ZX9Z|n@轔9Ηܼ۟NjONo~yeY5Zl~=eI#pPXGSv%DE\%p/MTfr8{Q9| M4NM%'F'Bp '\!3l\88Uu*1fBP0)؍Z ~ K?).ΛǔǠ܅avҧ-I)Wץc1/wf?)edB_&5AAd  b*ɎgVS0 F`*Rnwv.%wp MQT" nQҊ\ϬpFK"Z1ohuȼ ގyWtލsM0cP_@P1ܞrR *(~Er,̴Tdf̵y};|4YM|;uYde E/F&.L:7K?r?9hvX}1$)CGPn 'q!|#߽_!\@wypƘY6K+MDv ԜOudU:Aj%}6)i~p2Vor&9б"#{I9x?\B&=zt;ceexs= ,!ZeIMDBXRlؐ%P’B,Vr[5,JP’ql<bJ1ͽ4A-Lɖ |+J:@>I~]1S#ٱC=&S@uYoߋ\!p)9x!:!kNEv|#(_'huVnzI41Ey1.sYy5.kaAM㍛_ar;_ųUE^{/C5A]m }~14}m!.o^z|?^$0Pr( SVz_eߝOm\9a=VkĀĘ3(V%E:,Cv ^1-yԂE?O +p8\;0LF86l{ vғm lIՔM'%:[]' 8x{k/ʃNE`*i=ou.9 K |N D& UQC@b'CKޭSW!b4&RK3$J#D.(5FN-Ib8؏\RqVJy՚d"##JűbAP"X2.NT2eR&# .v-wy~nfl:8o$wGjhvX~{|3W /€Uߣɿɼ1ҋL \7|:M* Ԯ^Ǟ n;u}vd*܄m}K %D=z`.ԮoXJ.4D]Yܾ/%N͵ZQf]3@aSsNQNP[櫫lmۯ$Ѐg\4Fr0‹(3V~+D&ĐAP)z, g's+XI؅ )[O^=/:~Q +GHHDlϾ|>`H}SNrjt6+D3YM#;QdWcYKnNa=g*cW1CЌdyCGGGjށ[MDqXdtYH2!*Zgv%),?ڍ828"ֈ; dQIIldDiUcnBႧݰ#,!XA?vTgb@됋u*k?qrQ1աr}σ;za~%3\g7˳~;"}< >)θGU@;;) pO "i%kAȽڇW.Cb.&F٧gz(x9&TED42RUD$/%bxx~6J֯>sg3n ']$bAJ<Γ&ªGɔ, Ӊ0bDd{gғr\%ESʎ#ߒJ){aɖ$BR=;Sz\n!Np܁-L`3_ q%]v;8jڎ\J4k %d#()啋 ^z7L1~i?v^.Qz.sdGHy_Jw+#L#8{֟`Y1? 81WrUK%>Dd(oZfß.)}Yc1z=g}i},.jΖj,B`)ڕL^ E GQ/Q燩CD|he{ne4НO"O~%{ퟑ3xH^~9;cF>f p=pfwҞGɨ0YgDE _q,yjg,HQfU%?kdHFkLL$j, "Z$^j,.::' %`Du宕kQ"I$1 "w?մ[xdU\YkOM̨^xLKPcx:NɠŇbA6ǵ bI' [|=(ԨUa[>2Wә@#w~<ߑU 4?X?bX dCLָ'/6alb56̳j-"GY 4JSOF˛|2v5J|YcNp*4Ei$/)$>FX"v\RhnV;e2/RXPEGbr($pE5İ(ͩ= sOGxxѡ`B%%#UMʟ*Џτ0lAh ͗lN ;mCr)u[6-L]Ufcn,PjTBHbU  ala2`Y(Waqzd1U֖L Ee IQ&$egh9ގSc5i%K ^+G,eMN;JVgPhYQKT.P#4 lUD=Tڰ|g<8mUY3ܳ بyG+"BYtdu@&JK!D:}P!!zCfW<#++!Da_] Blڲ^<ˮp?,c+5&gCJ5ɱ;6R;32(^6v}ӖB<ǢgFe%B߹h?ƿ'B{hQ\<1n],\A_<<4xQ;lwvjfX5 sy$ctZf5 b. [R2÷֯*SFQo;}hs(zǤ} SlGo(e]gXul#k@I煗 SjfA[J'bSc21kE1n|rvYdjxk8Y:<)Y9Ts|ՒZsHxkh1J b5! U^ln*Z=D`Ghe V`"ACL#CvvGCo.01URsE N ZڔхE^nˀG&VcwBf-;PM@I`fS UKE{!;PG"f`@D9`_:Y"qWxWށ}{3+FC#N[Jdq"''~vB䕡stƜhK 5DAe)KIhݷ'2<%\_7=3wДj<H^BMJF2rΈ@ԛYMcAK ޷f2TVSAk :R\Θ,uxe\#Ɲ )tX i}io)ٲx%ڳzH1qؔ۷sh2 f $ baNj~Y-x5bo YdL:Qt-TdmqWU:4ZCכm'3yDVqIq^LRDoUМ:૫QadSъ#󂗒K'e,b\*6u@[LǃosUV7>b3l*hd-^ &&V% [Mƥs;Ia4lU:#Y)YQI:D+N-yTb(D>ᅞj;oa~hong:U`KUZe Չr-L-gEC"1-|>a)KYo)S !k\Dj UD%Ry# ?]<#ĸU!2MYh]m7̊&,Jx.@"IƘ])'ycPIP~^8޴GҊ$[zSKҷ[77Է'-nTY=`0Xu挰O2KeF R#5elXcO<3,E+EN<5(x"'_<^ӧ9{p> 웧fƀNt~FY+314L ;fsT1ơu-T7r)΋]P-CA<ޑs.ZSkz`3%MzҠ 簯K4e;8Tftwy[TIDRm@H_%hK&J"K"R^D*e^G؇ >?>;zW|hbdց'ekE CJ@ 3'C-:ÎR[!\a4;Swr&N¼2LF#.B$iH;CnP|? Fkr|sR&$R_W>bpg w YKj,.pc9seR2.]t@iOʮ8<1eRm3PF~4?vɖϴerK{8DPZ@AF,i\ci+G t @ ԏ;ݨƙ<;i<PCUY*EuY^#cᛵ+NZ<%\NQZ1z 9|Lb:i/蹐|( dʡ. `ɳ {P i|<5zZ<&m*ʼ"h?=#tw wƄbmLWog4hMyzg4=yh!G%~#_<ZeƉ[K$O< 6 Eo %Ŋ=7h;yEr|MAU% !A*Զ 3/F FzFFhxU,++e b.KLG w' {mg]DW2su2e7uGA\d )}/ c0|P߱$ޜJ űnؔA^C1OO7h>fg/o׫/*J3bSHzNKD ]D#=BK+j#ڔ:, Q''I/m(m"Ac[Ç79l>eި \rŀf,]d%ΚD\Yrf@o[ԃpO.nve09gVzR$R 6P0 T_+!p{ZiE "ATH{ś<VUKOT1XOV -NKɈ*F1nQBLp? sw9"{Cd~u$Nk%A#zSw4Sw4~':~R jMC 2HH6;0?=`bek@wDJֈP%_* ԮЮH=} B%ZB/漑4UgC7%o,]paځY[ʫQ8{? aGeΒr|V8$ZP^@j63ڌU&7GG'*CiLb8 m6p":j3t4FS2c@Eo1tÿUYia3+0Bʒ /MD(8,UB%VRYF(" "4PLTv묏$}kl ,,24+ox8y!$x e䳠$-`fF+QňG_@&8 (`8-WĂ^YШO[$xyPzC6*LFWjGeQ7* ـ^ZTkO7hTEpm&~胍ޓ瀶%l B, Pa1! ,`G7C["Z怑xd8a9wIͧ AB # UK} wWW aQW0M. Jn.Ldʮ-L Dٽp ʅ/W1-+zm}׾>rJw(*~xvuPp6X?zޠ=JG}<ϓ3kbnrOvZ Mgބ'|fk|@K--EcT+c$xNٯtՏ>ۇef/Oyʈ?)FUFUFUFռ~qZ}.}]:?IygȘrN_ 2߿&t|Fg7@* yF!,&L Pe0V utJl T1z9PpLh”J{}d--M?MxcˠrFsX㼸}ᖐ)'qw bz5F͎ r.ቜGΐ|leK ~vm .%੻EMΉHj=qL&(MjzvSK9 m^޶)@8Vn\TDdQ5{dU<)y"C O"Z I.'ШT+ḣ@a T ݀[U-C6\0``Yّ[^eh(W`(Ky&9^iNwY;VH;j3.hئėRuN?9ᘕ^="xLhN9vG,Y~I(FF'"դPr =QCZ@ҦΠ%JF`Y6'n,D(~( wt~Q|.td댞Q|QRȦ-8Sv9gIhG뵹~湂[{ Pʅ=xT;$ w,W eYlO/ފ~l ;2l| jhwؠr%UPAQfޜj nhw^\׾[wM)1{J xvj(:RYZM DZҒS޹6*] )XݿF~k\ȵr~OHGU!rU\"WM"44JLp5F-KPȉ(hTxG]-&*G(|-\ъ9bE]F3n߆OE%$ru7q!Z2*J6M') K(M*.D|ѺDRr`F>2 Sm^϶6lgL`㔑w 5F2m)e O<3yϹ^Q9^C˰i6U  4}fZywjCx̎H-ȘHNRi/Sέ%.#[g_j *Cse dIޙ >X3e _%|vh:pT"׶Q;{nJsXjĸ0aF)|HZS{>_YpdR|5K̑1Hd;(tD+ɓnR (\Q\0 yng3m)*gped"SStoI;y#hˀ۴0, jZkI&QVkT !}܋( -N7@ A:\\79 ;" )>Y"gvܕdoj9 I3zOUf˜/e p AEoħ݀3O=HvgXw%Uf7`2]YNBc}(2\ÎXˊՏ^"P;7A;/DieA;nW @x{ J7ڵ Q@][P Q:A\:k)ұ:T'm<~ӹݏ؅\BiKR]$Vk΂*[߁]ۻ!'5v"?n$F`*r$e12caKkR[l&ޅv|6pjs{OZs{DZ(L;)J\;N{K)ݴ'ֵ?ƻS_Wt}O-]Uױ]JZJLJBKjRR-0fwhR ݀:'كl2KCN7BnНUCiW\9oM0~:"UCjlBID&(\0uFrC)c'dOO1J"([# 0xQbD[b3(Rk | 1EоAODֶHBi)ƴ^EHSw\+Z5 xc"m(F ZWނ?(F7~PӍ vV B!lf)"$S(𑙂,.+/@cO4ns`ʐA7Ll H9'h"7x Pڡ%DMj٤@:nDzVd{9|uB,IG(fK=7_`\R >()r9GH}JJwf78` ,f7XFhf7hȥ}:f7\@} \>\(f7)r r >k@gpPu4-sBFڱuLTqdFym]2y$mvFt71l) \0BS.9 tNSf/_dRtsyKdzr>I:Ψvd$e*[lATCEƛe2u(,͢4r^8V0(103-B.֑á|lR j͒>F H9+:%GD9W !a0|M`7g11F^BCFJPk3 ӵFC1\C?,bgI ӳ;e}4nכ5s,D8z!CowZ///NXM "8Lj(Af@v 9E)|`SvN!KK٧'2HɔMoE_9sW_4SmX'"̑Ye- 9do9e _/Gߊyɷ,D!P E|:Mj`%kbTB]kؗ)8$i+7])[uKRyptWulkkݒkxpt FwXtk=xWc1}ptPLC뀝H9\7k^: ysEs)ٔ8=r-Q&}d2 B<$+ƧmM%M|FqgԓI O^CswnM{nC^5(Akp|K!@GVFn ֐ϋzy2Z, $)1r6TZ&-HɳKҔeY2*?q{:hD'thU2;@7#s0 4$,RdjEjYG Gy=h9䲔A\jK\d 4/cIO-.$!2R(8OT1'l+4jG!#oþױ0DkO]=Wl'a )ޥtTa;)UHRj;!V}tάmK>J)vR k}z[yp@V^\qOy?} ?χgخʈ3l7r~/b-7kRnTAG"/$~R$~׶oĥYGc 'XQ:deƬ/~B6'0cuW5olBf~5˺lzz5}|Tq/,iN'ĔʯIgg߆bouTK]28NrH)B&ymF><$,#max_WB#ҭO|B}kTeњ΢=3g5gP[q..v,gv ,\7CƮw˛wCݽ\?}$tsQ`TPh~L)oGk29B/|Xۉ:ЎNZ-{ފ'ZpIhAzLua֙[V˯uxocŸ$C~<~rՏ7;oLT.%ٻm$Wm_D:[ںŕMSR.mmI'M߯ɜ8$G]#xh xH]2iZn*'@C Q^:/ 5V_ҵ9/kq >jͪeF2%Lk1Jͅ7%PUZ6+f]ز¿o|8%9!YһecF5f%>^+% Oꎫx˧Nf.opRۋPٻpNC_nML(((UqΫkx[xDr,.j$ :# hcڳ\vH IA=l9K &xP+J)Jx;P*+%QA冯$t ÅRۇ˽#;Wu |z1j\wMF?) h!? ZM ~85&>=BCDŽ7ʿK&x{Nv tm߉x/~= RMم}=Ҝ($(0LJў, (6z4Zb%Kb=Uɰ1BѬ>]r6*rk.$%D $Ԁa@tRU+3 N.hVh\`/~ee٭R]գnobpieQr緟$.sՠIɃ?~=Є6T+VЃwgߞik"׼`nc*7hGJMf얼& ޼rٞ>I;qʅz%3״Tvy櫋-=>l7kު^M`V}XTCWpmvyvw9~!޴ocjOnb|.ן\ )H[5ܛBpԓՒO3xT1A M_wn˦u@?lZi6&pƞ5]C l#]j5.{yc'K-a6WN~9>;syvE<ېZ .uܑŽm Ye~溌?\ĠC\+5ݖc\5L#!U.?'nN?}Bءظ G'UNad|e=.$?~>W'1pޫh=1\$znY689:ՙoӻ)r1H1wtnE3Q%[M4˦|Ĺ4ûbb:\EhҸk-CwB^ٔL)YAVFsz{li|p&K ̔,߽yqBsQ9ǥ>[r^ `YPfTE}7kƴR &TIwP6~0P5EL,j~_J<Rm Rm;D4PEVl 7i?}Vil]R)IKm7ww -*i,dRIw<6oe)ENLG-29JTJ]PG˱]6J He2ٽ@hO|k2f@I(-RAiUc(NiI2<58v_+(}I*z09UF{Y> s L Հ {f|!$T/ Ur䅶HVRkQkZKR3@@Z*UMU}VwZjd)uH ҕn;*,#k5S'##@U!Z*-aJJ@d \ETq "Z䄔01.F+0,Qʂ!V;KtRK*=ZT*2KMZP'b )m0 ˲;ݗu3z辬w2B 9Cbj:WNyN)xLjg`Џ1HUy ڜ*CE8CgnC;(3ܙ2 1d:6ȝ6 x$ќ͌H#Ϭa*u5iiܙ -Jh/5")M{X֨8_/ځɇ`KZ4t\m9>Ŵ$ a0Z2Mmw?}C=tNEsQ݆;[HjwE] .wHOo̾?:B.j/gk,$c@l)wdumGٷ ǺӑB$ + [(#|vHHE;f7/,rմ9s1r& S5_3Ahes;O[7"Ӧ%7[EvƑȓNI\<='F &62y^x;OC#v'q -`Wp䓕F{I5 + BA=`ZSAXYEg{֨37@`drvo+Q|q,]qv^9CjMAd2f!.Iɏ UmD\ R[rхahD= C}X 7,ڕvioûŕ[.).mPFkF`-0ѻa!/Dl #MlܷN;x1W@ Om$z>,䅛MD쐔2m1vȔZ!H||;^r_Cへy,K辚@SK^m'INfߞpQx_,<GOGx0SD+,*mms{1׳'Nk>~8G)oBuw]mWձfLar=#9赶DQ>Uԅ:!I KFv\(hf.beBxZV%wԥ:g]JKMe4ZP=a8L; !:L"Qy5|wj!;xF DNhLRj~T'.ٮE:M~.ܸwӯv7\G^8"*,PFLK36s&uA|GWqh/}  Wk`eW6tD(JZ\j S!WM %P T1`+O+NF|\LТ QI;M}g4GP.TKXjyڽRd~@?{Y>d۰^uya7\aJ*MSc;Qx)h oq5 Obä1&})BrZ"Z!x5—J%h;8`t8_AF5~_GGݰs>C4%F~Ir# 7CH 7[\V:ӥ cYweYYQQVs"eJj{-ބQ$( !(P<`d<8FseI%󢤔T33BqNR07c7`D`c9owj `=ֻ%gyFhJ1M$$JR-QZ ՄXV+'ρ C?l?-&FH_SKƎYaLQP(|? Y}툆teRSAoiOit\Ho t[5*`:D|D]ρRcP- p *ʸ78"e)GYRaBIRsĎ#ǹKt^;NAXh( Df&ޔH>a]j(u\Q7:#acGԃJऺ' H+@F`8:w0gA;!kC%;$;vuim ڀv݀K.OK䤇.-ho|N P΀ vE y |wG[̆S*`w[M`N]^n&QƉhЎMczͼ`8-fnӚ,䅛hMZ4`Ĺt#ѻbb:\E(-_)FwB^ٔL3Gc)-rPiJ4)i[^MX}\xLֳuYA^iv*`{#]hQIcozl3uVc*eNE+Xg4.P)=vAJi76 ]Iv>&uq #5TS7O%WnVsձjRx)-/-ۆb,5O2$ cztmI j)>h9$l-<\)25ث?0r:y/yPQ\OLd- u2sP!">Tb*lz]2l\o1j{ jaʱ5ae,5ƭcN)B 6ZpV* \ = `=~w;?hͱVտi~;̫5sOB)e7/VT-RX׋qXmȚ]ۛ8Dha5?~qD6t!EoxC |J !)]RgME*gJ-~Һĩ0%Sijc=E@unռM"V|@?.êo]!PJ~XջDYA'yU ]tpi+rP4`qmdESQPHnX*-dH K EMHLj1"v(d`/ g-h/BZ)n' Jc `$5 wem-zL0M <$AѫL"iO5II%a.mX _UuUuw-|۪BT=> f8-#H[&r  u٬0 $!1/u %\XSekGh6*3و8/7s{ '۾SY5e]-v]vw[ڑ\bW)Zw-?bq%8uY2s.Ttqiq0:)ՊW:fjSKq FUSa(;.=C'%1RJ%TN%7>b!Sv&{uuF9ZSTkoE˖]o[tϲ(kPFc7VKT-/х"Z%S |$,[% DSFݪJ[L¨7{+mڒ%{}JVw Ku#Ʊ.TZ:r"Ω)پ\eyAӈjH+9Q Α䊉T zd1a;k^Gjג[`B,!( G9Xu<.W+i+^Ծ1z;CNyH"4guWDG [0%^>DTYIΗڑ)5[:ΧiICbWIvfVSjԆ}+F-=o1ܜBd<.eQ1jڨȂ}i8>=50NjiPrQ6rө T)VKѩ}Gv: x<;v9Qݺ_\DdJwk79vAԾv;xjmPavBB~qݒª;,ExIYeί%e<ᅤ,.܌ pr)2b9rxgq#mr>)J+,1<.F;[ ͍+m̹dmYlcڤd1LorUtޒުjwzr&zmܗƧ]lmϪ&Zo@W & >]!z|FaM&zC 7S*t+9w1ϑxאLL#EZ $3`Y$;U-fTSCCՍ n%/aHߕWGhiMHgC'8x[^VNg1KBDnaIs#)bp q^P-8 ^K¤c8a-ȸX722qT!U *=1`DrAj9OKts9eJ&ͅT@nDx2IPGD]H"6LR1 i :$cCAppDE2mUB]Ewj\P^kjw=;[hqOA~`i ̗VBZq]&+:z0Jb# )!bv@!+d?]mʚib=~С-r4P k!ǥ'st@k%O/tGd{S 2Nנ\YƏ#Nkqi}2 .́.۴=woX:Fx]>|oPެB D\cz p9k׮) b)DE棡?拘?AɃAfl8AT1옲>RHB ‚+X'M»(ӘYŭ$lbx\|p),p6־:c_֙?la~|p\&1Q^ۢQ2K4j I! I`C "ME 262XibPXX6 ' K,JL \w[Qwo "ӗ6/h)dwSh ,Pͭf=NPnj)f;JT3~lpg`-m Au/U2R$ǑftB9uq&HqJzUGWoT <pa9qLrsseTave)9?- k[R \BIGAS#=`w 30}X5?@i!PTLs 05h#kV {u/ c% Dd$WXpzp5e%ZuK5glF0rIi;<xQNagn> G$sQXJ%x EpEn #\) xB=@ͻzj1n<؀|E(0Zng)j#P,?F\Kޱٱ!_/N8KM @0rSyz\&:_x ZB\=z816r&,bJTtC-s 5%͗hK<Eɣt~YI"'E67?ܾ+z*4^k6ܣKlLd+SD$ғĘjRFyrxq|L eƄƷ\ÆB'D(cMO5 8[9.; 4% `G%e ." L+~x5LB>S@K CeIs !jLTk&"{r-m$pEu7ͧ 3'@cEE?\Mƈ{VW%A$EXEKD"oljeֲN)cրbkUJDiM`;iuj4t٢hNNǃB)-.ⱻR׶KJ{=ja#lf^(,^("8݈Z#7֒xy3^ * j]ȾY Y2.$Bʸ7kA.R&>l:(aL6gǁRg")?;᮱| >حG Ow(fm ZVj3c شW3Х^g%6:,%Apu#yi 5XU~g8'uvOo~ܮrF(e/nKnef_3XN)l')]k<<)q gx݃7%rϗ'-~&r[5e[Βe׳jY5 Hp{#zo6685m*V r`X5^bRf)|'H`)B{15SFYv-k]>OF-xW?&9(@Ju8Ӭ}2)}}ԟZ .*LS͜ SiXȒbIE7j(Bex{R!azɄ+}d!%0,r!RMIP{vn+p0svhuxqax[E{h Dq)2 +֎% }_K'A6֮ϗ`pC6Rmx@2BPc/tx#Vxƒ34%xreζ"ZЀMǔ =w$AZ i=NQywh)t*l@AEhGVb1[hW3k·c 4 3!Bs>pe7?}[ c!&g"V~v GxkgJ|oۂ&ǣ)$`8)*ǼȀo(ʹ8;d“Jq5ȕ8%+&UJZ+^]EʷChG>͠0I u#nQ+njZ\bu a#V bC|FL6%O)(T+u2Ca<u,́րn*($Kn*KMոp`FV-rF3*z |Ȳ2 h89q⋑D@8T72#GfaK6nu/{ |Z~xO$"Qy=)6eҹv2j]CK_g-%,*+SDP#jlJ\݉.!& e$%@y'JH ?ɽi@tk۫IW1M@! '%qMA2494^OV翆qס; D!"9QK(YB)Y}΃Z5VD7LYCY1DpQ]N( QA^@Ep8:=!SC G404I#rD($$re.Ʃ 8N.ӹ?̡EwD2h.G>FPku1<X拦u>ɖ% ̚$'XlAJ ;1i×{`)RzN|(.(d P#UIさofyXZW/9y h64N|^5vy23}Vbnt#] M< iDF%UrO{=/evBY^,2FلYb  ǩWJE[?m~9[qYQby# Z6z!)Θa :6\r-OerV)YW~j_}`F UkKb4hЛ*u2፪^0N8VHo8:?y [ns{w._b/}B)~hά/e~sݾ]|O̬hBfojd8QxT}$t0~2]=h5?0]@hi|̆V" 1r*XyUN'q^/8:<*?; ͩo8W6vPZٜ3l_~,Ad ra|=P;_6oM;5 W[]0|KXl4|r~@M(DQIZB hqp;MRazŒ[Bn?~暯k6ﺫE$9d" 6 hpkǓ;V`u|ɫTZ\(`$  ؋NY7ǏN1v9l)Q);Mr{lK4;'~ ܾZ;)U{=Aoʱߓ8/B2}ʜ;-ܞV(B]d".SZ~ 3;.\ޞxt Sa` {t0߃E.|aQڅo q-ϽׂKˏ65t4%R\Y,ĥ 7gW+ʴ#;mҟoW:bn=.w17H)pqd݅hZIߟƖZF8ߵKab w]ٻ޶r$W~b Cc:fz^v?m%dKr}ݎr^[\U)ك\񻳶u#0܅n1.B7|<p: 1dUХ*@,KΆb/VJ1TJ(%AUw(W}t wPbm75\1aMRRd}Mr!Fz!"#B/i[o^f t߱zҊ7˥2o);PY?$z(- V)QK؋}}vq'6 wjJ΋w'Hڥj* v7^w^7toIq:H|K{&Qi g:Vj] ѶP},W=wDϰ]h^}; 9>I$[X?GhI^!G:H}A>IW5'#tyE$=G)>{/Hͻ31:e4I^ԍSw- GF[!@9tl~N7|g?n*ދ^F{QXĥgr7Ý Tq*zqdݬp0DT>e&Bޘj:eϦ"=$p0.|ܮwpOT%*p)\rnlbGpGF,'|l&)(;IZg9͔fw"bN$6?yހ5h{Ѐܮ一ުi9y$W+S@#$,6(-2oF>ؐ8 ۗIK3kJk?R؞#1>ڴiΟyϬ4I~k5tG^5JaފŎXt4Kc΀-7Fέҭ.ֳN'H1:a+pe+Yx@t?رQKG1vYVe6Lx4@L_eg,,Mt0]188;;y,Ri sVWc>ԱtFk8:.>PytB`7u9GaN3Xjڌ $V/Ag1I"cYoN4Iﭑ8ORNN+[|sx3ɼ.`n4Xvzmer^_:rYSa>rA\^/0hFR *ք@y-j*D\dy \6LP"L<.%:;\Xr\CX9͈hLiX^\$\ GoguV#N{uCv{yfU<w[G:1gw1Χ?`h$F 2 `O9a9`OA7]V}zXxeJn}տGyBqWu֞SA}lWHPvYK7QjDK^|P'`AkSRĠJ"(ڐPn(5Ypw1tlExMQ˨ {}ԂU8I~[W,ˉ\ $^W%CBQ1NζT*yP^`bG沌< GQRF:fGs !<{{#v1=[K{ZkZIḵTnZ}ukjQ2j7-=j-5MK v]ļ(1'FFa1P"Py?7X{6T CzozMYo/q|'xv10O Q'IAqG[T~%3]ȋаDdL'O"8eo54{<9!l+ri-(l%Fʠ+'bS_NHmB/ъ£#aB`Xp;?K{ݞR^$C' RQ;ğ~b0& ~9P[tn]?DAdKiarfD0!d '&o ,RGGU I}e=q9/1a,/GPN 'v=Z;eAĒ%rE< RQ(d<)J-t*bA~On"8/>}XMjJTB״q.v%g`Qk'RyOxRpcaI,T[*,OtRڥ[R\9L9fF 9e c>C EAy,Yx<#Wy t@'e݈Ӑw[~rTk%Dr'GNMl!ZHBYoSk-ZtakdۡpۭG{og_6澵'Oޢk};([v z6 ޳?lyTm(sJ"ç"^#TXSps\ gP#0E:>aY6nMo)^M ~l|F|Kvtݘ3/X.Mʨz"Gݚq>81h 'TZ*;ņzXQՆu쭃hN?:i$=P9~%0CvP Ӈ$)]$EbGGj!Z҆"娇/(Ʃ/.gҊ ;3_9KlRARԡL``E~4v+Gs>IEk|UWxM`١5U5pjVH"'bV좱uUZI,`Lف]&eq݈tuq+x/iW7l}='l8c0jAUr GCk|3>$y݊lMM= SPSy5[v,9UXSv/'9:H l7owF)v(#x9(e0p=[l]V'Rn{9FqqB2/|M:@'Nt.Jv0 .xR-D8+CሧH8'̋@ 8uKP_W˿ʖFK1r+G|g-!p< O6C],-+i2wxo($߄CrTAzR4OG^aRhokJ[f^<|n5V环Yck2j㮱ꡥ}#:%zOG@ӤiKS;A zcqچW9-zN qY|#,%,IY`dYdqY`ќx+'E9ښΓ_bA'^)Пϝvx{_tٽ%:|юȦ!`M0RtxWcuRwZl4Ut{vNMCZSbn :wblw=3^9ZrKbqnOVz~91eB:f޼_L>cη-4RÖBXJ,my`# 9׭*:.ygLJj}ؐ7z ~9ͥdAPBd+تhM Dž/dPZͦc^Rl)g)sFu&٢XWv-|1)\ѵ?ŶQ7/$Dc{`G+ͨnfкCEpыL7쫢F@!w1G{ z{5:bߊ(tlf[>Voc7 :%i1jݴTvTJahݴt3j븵nZj6}RiiukHE6[wfǨhi)M -kj)nZZϲS׮ϷVݣ~dԭ/=n-ղjQ]~t-M~~|l9g[MEsruҡ#p*kQ-9o%?{ܸew%pˌ/f*O"pVL_emHɖJlh4ЍNj]5ka{}Q\~^-9ŷ4 s tJj],oon&C_B$SL!m$,%.ShS2'*( 2^8\0焛-i@Mҡlo4.8uTs5 LjJ$%qԳ1/)3lj S+@P&% KXP"3QB0@ύާ(*`\SxbqZ//.^|։و@aKHd>o1Eu"ՎwBi5>=ȿKRX6Ҭem;7ˏPSA`v P< h݇qs1& “F?4/V<]{:ě ףR N$XG|pThŨlZc&4][*x8M{< 簢b ҏޭA]h "ګ?hz_`AN(f/{f0˳bDXMN5. ck`Ɛܮ͖Kqxz+*7F.d.7\w 7ۯ@⑕dfWh`}iwvSifn_׻ ?HSFgT m$&6E OwA ˥Q7y a׉ 0Ըe1hPM:#DPRѡTKW Uʚq˟}pz}k5e&D-ԣ-RMA cÕ(݌V'# * np.F{/oq--U ˔$Rb h)(( 흛K4_e*FlM'~ʽnZְz,LBb(XLlE3q0byȗJ4gaH"&zF4{I,I~taL,&  _X*Cr |&<mQE' 4I 3k ;yCfn KX!/yEZ%y Qm^ qЅIJ#8$ J',>*p`yI:=>/0C]!|(5_фn i5'q|m_:SA7:X7Oui_ѣzwZ EKo1#Bl^EϳDLԪ,e46aPCN3*qN̰bN~D3SnULh(PpXfIRE){ ו p;9z% w/_1fE>]I/76/iϓ"ivJ?%pTzQhN59D;H*9)~~nT<|Dc2 NfДhiT^Җ)-;LC/6Ɇ}l=ok%)1Vax&`e.9JXȺ<>dlmr'ء۝,`gb 't&-cndCa:-H]iQnrDխ>L)I&!:A1,Ȭ\MƔ#D+M9a1uCM$2R,aD[i(o07xn>A0w9;OS}Y &Ow01$b u'&q>7  Z~JBB:mR\| gAc>ڭ1IB6ԀeAjބ  <m'Ղi^:&ZDo '9Dșc0Ef%K|3P~|b[V"ޏ'|3ii/oe [(](oD߇r,:Wu'8RЕ>7\J@6jp|IZhDKCκf3G:,GzZ39iR0Z1T u'5sLZ'hO5E!QFi*=R]F]Ų=@oB. :b1M,x:=]IАW1:% ?5z˺ A\ m;\R_Оh+W(=mF|,!&m]?-uBC^Vt*z7!<] ;k~P57H¦i\6dFST_nQ)#3bi`N&U0E 0-ͩ@IKQKy=ZNvA Zh#Z$V[pb{cǵ5儁3i,"ceb Ղ̦NQq9ˎjRЏWD[xo3f4S/GF2㘝PIDgi61+i*R~] th>/y)Xu8"(+)B ]ڳ5BŽ*RD[QMsJfi(Ud)Bi&R*E6' ?k*2A`[RQ3R'bO{@5щJ T 9/,̠M$Z+$LIC%eJ()Ȃ LE?#hS`,M0 ՘QT2F^8ϯY^-Bul3T4ba25OJy 5`0f{,b|TO{zBl"X(:[_3L?\IΊ >^%ףm>qPo #rcJ8&vBgOaΑQ4e!by*ZlS)HR ɭ;Zb9NaTf pf3I:eARb2ft%LhE[M!PjOEIs3C ^ufhF_>sJ ;>h!/.kVw&VQT)|ނ#$R}ڞjJrx5aaQu z}=v1iTЛޱ"Q^@zMUnc ߤyC4)E޶I nQ<ݻ*&$E9M0PdS '҇˺ܢHuH J'FQuv^V{}:Zݳ y*SƶuӥZX BT'MۀR޵Hr࿲%@I)=a^ux${Hz\g)?Y^ư7"QjƳ&dVѻ)Fw[:T[-'z1,MtM1J>&;.kTeM6\@mcu8;r]WB#; [6mlO۲p3X;~yX #pOܽ {? 4 C=\jѶG2ߞbpu"O|ƨw?)~[ 9SI<>%4pfðRiW -KS&Chh}% UJ6h[mŽs 2aSym#wzؒwT@|Vŝ4$,,*z4jtZ9c ME ?LJ_N$º<诟HCiљ#Fǰc _)iT@tUM Rl#%4[h׼un.|㡸1ƴǏs!w4}m>,p\{z_Е]I,O Ps[lЌ~VŕžY MMݭe%^qRu7gvq8͆@|+8уD KD1iA46:OJG^h k,{j-(RT.nyg6|Q䶟Z~9-^cY@9 5bL+ygT¼J)f.C87*ʊ"=H|E֥]y\MH18ωk.zgE_T̺tp@j,`hYxV%܌<3Çl-MC33H\F/ggp?L7*&Vm4T-ɦ,U o\CY%{ )yxl|bԊpDI :Y+}4Id*Ȣy#쌻 Vg}.y3-~{$h¹{ /:}b5&dh9]&nI~2r(ۂ2J}@M-1PxǶT]Y؇} &LU۹ʂ^ưQ\ȃD\5vNb0 !*b#';U89Xw]@(ԂGSw-U@Tnk+Kwe(} AcwovP(}SPD>ǟow~1*)bE([Zl thІM^{,Q쉸uXI|CdAeJutf[<7ywt+Da1]Sk WA#\ʵH45**X4%(21عi/z֠h_ 0a)c(~iKT`IYjE[&ydxiw_J 8JݫӸxSLd Z6鴴6CK9H_LM0|O6!-mʖ'[vjFV a<x)ml"U4 ,obH%0#ՠB|QR󘋅13"ح>moAbc O t鼿в-|ک)o]Хvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005135013515134142261017677 0ustar rootrootJan 21 11:10:56 crc systemd[1]: Starting Kubernetes Kubelet... Jan 21 11:10:56 crc restorecon[4561]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.891767 4824 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895214 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895230 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895235 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895240 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895244 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895248 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895252 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895255 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895259 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895270 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895273 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895276 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895280 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895283 4824 feature_gate.go:330] unrecognized feature gate: Example Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895286 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895289 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895292 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895296 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895299 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895303 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895306 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895310 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895313 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895317 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895320 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895324 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895328 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895332 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895336 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895339 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895343 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895346 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895349 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895352 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895357 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895362 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895367 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895371 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895376 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895380 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895384 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895389 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895393 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895398 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895402 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895406 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895410 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895414 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895427 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895431 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895435 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895438 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895441 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895445 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895448 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895452 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895455 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895458 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895462 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895466 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895469 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895472 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895475 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895478 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895481 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895484 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895488 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895491 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895494 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895498 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895504 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895869 4824 flags.go:64] FLAG: --address="0.0.0.0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895880 4824 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895886 4824 flags.go:64] FLAG: --anonymous-auth="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895891 4824 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895895 4824 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895899 4824 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895904 4824 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895908 4824 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895912 4824 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895915 4824 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895920 4824 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895923 4824 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895927 4824 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895932 4824 flags.go:64] FLAG: --cgroup-root="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895936 4824 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895940 4824 flags.go:64] FLAG: --client-ca-file="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895943 4824 flags.go:64] FLAG: --cloud-config="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895947 4824 flags.go:64] FLAG: --cloud-provider="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895951 4824 flags.go:64] FLAG: --cluster-dns="[]" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895971 4824 flags.go:64] FLAG: --cluster-domain="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895975 4824 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895979 4824 flags.go:64] FLAG: --config-dir="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895983 4824 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895987 4824 flags.go:64] FLAG: --container-log-max-files="5" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895991 4824 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895995 4824 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895998 4824 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896002 4824 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896006 4824 flags.go:64] FLAG: --contention-profiling="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896010 4824 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896013 4824 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896017 4824 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896021 4824 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896025 4824 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896029 4824 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896033 4824 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896036 4824 flags.go:64] FLAG: --enable-load-reader="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896040 4824 flags.go:64] FLAG: --enable-server="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896043 4824 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896048 4824 flags.go:64] FLAG: --event-burst="100" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896052 4824 flags.go:64] FLAG: --event-qps="50" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896055 4824 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896059 4824 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896065 4824 flags.go:64] FLAG: --eviction-hard="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896071 4824 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896074 4824 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896078 4824 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896082 4824 flags.go:64] FLAG: --eviction-soft="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896085 4824 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896089 4824 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896093 4824 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896096 4824 flags.go:64] FLAG: --experimental-mounter-path="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896100 4824 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896104 4824 flags.go:64] FLAG: --fail-swap-on="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896107 4824 flags.go:64] FLAG: --feature-gates="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896112 4824 flags.go:64] FLAG: --file-check-frequency="20s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896116 4824 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896121 4824 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896125 4824 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896129 4824 flags.go:64] FLAG: --healthz-port="10248" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896133 4824 flags.go:64] FLAG: --help="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896137 4824 flags.go:64] FLAG: --hostname-override="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896140 4824 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896144 4824 flags.go:64] FLAG: --http-check-frequency="20s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896147 4824 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896151 4824 flags.go:64] FLAG: --image-credential-provider-config="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896154 4824 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896158 4824 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896162 4824 flags.go:64] FLAG: --image-service-endpoint="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896165 4824 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896169 4824 flags.go:64] FLAG: --kube-api-burst="100" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896173 4824 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896176 4824 flags.go:64] FLAG: --kube-api-qps="50" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896180 4824 flags.go:64] FLAG: --kube-reserved="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896183 4824 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896187 4824 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896191 4824 flags.go:64] FLAG: --kubelet-cgroups="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896194 4824 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896198 4824 flags.go:64] FLAG: --lock-file="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896201 4824 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896205 4824 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896209 4824 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896214 4824 flags.go:64] FLAG: --log-json-split-stream="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896218 4824 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896222 4824 flags.go:64] FLAG: --log-text-split-stream="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896226 4824 flags.go:64] FLAG: --logging-format="text" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896230 4824 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896234 4824 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896237 4824 flags.go:64] FLAG: --manifest-url="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896241 4824 flags.go:64] FLAG: --manifest-url-header="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896245 4824 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896249 4824 flags.go:64] FLAG: --max-open-files="1000000" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896253 4824 flags.go:64] FLAG: --max-pods="110" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896257 4824 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896260 4824 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896264 4824 flags.go:64] FLAG: --memory-manager-policy="None" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896267 4824 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896271 4824 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896275 4824 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896278 4824 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896286 4824 flags.go:64] FLAG: --node-status-max-images="50" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896290 4824 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896293 4824 flags.go:64] FLAG: --oom-score-adj="-999" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896297 4824 flags.go:64] FLAG: --pod-cidr="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896301 4824 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896307 4824 flags.go:64] FLAG: --pod-manifest-path="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896310 4824 flags.go:64] FLAG: --pod-max-pids="-1" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896314 4824 flags.go:64] FLAG: --pods-per-core="0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896318 4824 flags.go:64] FLAG: --port="10250" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896321 4824 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896325 4824 flags.go:64] FLAG: --provider-id="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896328 4824 flags.go:64] FLAG: --qos-reserved="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896332 4824 flags.go:64] FLAG: --read-only-port="10255" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896336 4824 flags.go:64] FLAG: --register-node="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896340 4824 flags.go:64] FLAG: --register-schedulable="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896344 4824 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896350 4824 flags.go:64] FLAG: --registry-burst="10" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896354 4824 flags.go:64] FLAG: --registry-qps="5" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896358 4824 flags.go:64] FLAG: --reserved-cpus="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896361 4824 flags.go:64] FLAG: --reserved-memory="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896365 4824 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896369 4824 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896373 4824 flags.go:64] FLAG: --rotate-certificates="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896376 4824 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896380 4824 flags.go:64] FLAG: --runonce="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896383 4824 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896387 4824 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896391 4824 flags.go:64] FLAG: --seccomp-default="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896394 4824 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896398 4824 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896402 4824 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896405 4824 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896409 4824 flags.go:64] FLAG: --storage-driver-password="root" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896412 4824 flags.go:64] FLAG: --storage-driver-secure="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896425 4824 flags.go:64] FLAG: --storage-driver-table="stats" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896430 4824 flags.go:64] FLAG: --storage-driver-user="root" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896433 4824 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896437 4824 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896442 4824 flags.go:64] FLAG: --system-cgroups="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896446 4824 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896452 4824 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896455 4824 flags.go:64] FLAG: --tls-cert-file="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896459 4824 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896463 4824 flags.go:64] FLAG: --tls-min-version="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896467 4824 flags.go:64] FLAG: --tls-private-key-file="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896470 4824 flags.go:64] FLAG: --topology-manager-policy="none" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896474 4824 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896479 4824 flags.go:64] FLAG: --topology-manager-scope="container" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896484 4824 flags.go:64] FLAG: --v="2" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896489 4824 flags.go:64] FLAG: --version="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896494 4824 flags.go:64] FLAG: --vmodule="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896500 4824 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896504 4824 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896593 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896597 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896601 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896605 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896609 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896613 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896617 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896621 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896624 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896627 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896631 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896635 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896638 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896642 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896645 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896648 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896651 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896654 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896658 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896661 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896664 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896668 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896671 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896675 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896678 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896681 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896684 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896688 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896691 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896694 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896698 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896701 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896704 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896707 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896710 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896714 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896717 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896720 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896724 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896727 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896730 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896734 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896737 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896741 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896745 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896749 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896752 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896756 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896759 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896762 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896766 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896769 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896772 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896776 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896779 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896783 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896786 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896790 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896794 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896797 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896801 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896804 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896807 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896810 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896813 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896816 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896820 4824 feature_gate.go:330] unrecognized feature gate: Example Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896823 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896826 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896829 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896832 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896838 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.904297 4824 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.904318 4824 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904367 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904373 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904377 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904382 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904388 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904392 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904396 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904400 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904403 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904407 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904410 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904414 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904430 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904434 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904437 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904441 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904445 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904448 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904452 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904457 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904463 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904467 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904470 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904473 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904477 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904481 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904485 4824 feature_gate.go:330] unrecognized feature gate: Example Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904489 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904493 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904505 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904509 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904512 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904515 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904519 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904522 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904525 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904528 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904532 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904535 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904539 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904542 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904545 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904548 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904551 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904555 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904558 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904561 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904564 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904567 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904570 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904574 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904578 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904582 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904586 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904589 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904593 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904598 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904601 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904605 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904608 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904611 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904615 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904618 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904622 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904626 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904629 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904632 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904636 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904639 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904643 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904646 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.904651 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904745 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904751 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904755 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904758 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904762 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904766 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904769 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904773 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904776 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904780 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904783 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904786 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904790 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904793 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904797 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904800 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904804 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904808 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904812 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904817 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904821 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904824 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904828 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904832 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904835 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904839 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904843 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904846 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904850 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904853 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904856 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904860 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904863 4824 feature_gate.go:330] unrecognized feature gate: Example Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904866 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904869 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904872 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904875 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904879 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904883 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904886 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904889 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904893 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904896 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904899 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904902 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904905 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904909 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904913 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904917 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904921 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904926 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904930 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904934 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904937 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904941 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904944 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904947 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904951 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904968 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904971 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904974 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904978 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904981 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904984 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904987 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904990 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904993 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904997 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.905000 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.905004 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.905007 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.905011 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.905109 4824 server.go:940] "Client rotation is on, will bootstrap in background" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.907688 4824 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.907750 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.908759 4824 server.go:997] "Starting client certificate rotation" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.908783 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.908902 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-09 10:23:23.933533086 +0000 UTC Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.908949 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.919532 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.920955 4824 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.921722 4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.928329 4824 log.go:25] "Validated CRI v1 runtime API" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.944409 4824 log.go:25] "Validated CRI v1 image API" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.946069 4824 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.949559 4824 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-21-11-07-36-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.949588 4824 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:49 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm:{mountpoint:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm major:0 minor:42 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:50 fsType:tmpfs blockSize:0} overlay_0-43:{mountpoint:/var/lib/containers/storage/overlay/94b752e0a51c0134b00ddef6dc7a933a9d7c1d9bdc88a18dae4192a0d557d623/merged major:0 minor:43 fsType:overlay blockSize:0}] Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963034 4824 manager.go:217] Machine: {Timestamp:2026-01-21 11:10:57.961268161 +0000 UTC m=+0.254297453 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2445404 MemoryCapacity:33654116352 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:d00624f5-786f-41dc-bdc1-a1875f167d65 BootID:695c676c-720c-4921-9e71-56b96f2fb46b Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:49 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:50 Capacity:1073741824 Type:vfs Inodes:4108168 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827056128 Type:vfs Inodes:4108168 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm DeviceMajor:0 DeviceMinor:42 Capacity:65536000 Type:vfs Inodes:4108168 HasInodes:true} {Device:overlay_0-43 DeviceMajor:0 DeviceMinor:43 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c5:e9:09 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:enp3s0 MacAddress:fa:16:3e:c5:e9:09 Speed:-1 Mtu:1500} {Name:enp7s0 MacAddress:fa:16:3e:ec:64:04 Speed:-1 Mtu:1440} {Name:enp7s0.20 MacAddress:52:54:00:ec:ff:0a Speed:-1 Mtu:1436} {Name:enp7s0.21 MacAddress:52:54:00:35:1c:7c Speed:-1 Mtu:1436} {Name:enp7s0.22 MacAddress:52:54:00:53:57:a0 Speed:-1 Mtu:1436} {Name:eth10 MacAddress:8e:05:62:5e:18:a1 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:aa:72:41:c4:00:59 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654116352 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:65536 Type:Data Level:1} {Id:0 Size:65536 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:65536 Type:Data Level:1} {Id:1 Size:65536 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:65536 Type:Data Level:1} {Id:10 Size:65536 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:65536 Type:Data Level:1} {Id:11 Size:65536 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:65536 Type:Data Level:1} {Id:2 Size:65536 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:65536 Type:Data Level:1} {Id:3 Size:65536 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:65536 Type:Data Level:1} {Id:4 Size:65536 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:65536 Type:Data Level:1} {Id:5 Size:65536 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:65536 Type:Data Level:1} {Id:6 Size:65536 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:65536 Type:Data Level:1} {Id:7 Size:65536 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:65536 Type:Data Level:1} {Id:8 Size:65536 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:65536 Type:Data Level:1} {Id:9 Size:65536 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963214 4824 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963291 4824 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963551 4824 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963704 4824 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963730 4824 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963883 4824 topology_manager.go:138] "Creating topology manager with none policy" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963891 4824 container_manager_linux.go:303] "Creating device plugin manager" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.964207 4824 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.964233 4824 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.964307 4824 state_mem.go:36] "Initialized new in-memory state store" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.964384 4824 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.966217 4824 kubelet.go:418] "Attempting to sync node with API server" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.966235 4824 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.966254 4824 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.966264 4824 kubelet.go:324] "Adding apiserver pod source" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.966274 4824 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.968458 4824 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.969090 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.969835 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.969844 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.969907 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.969911 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.970805 4824 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971742 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971764 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971771 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971778 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971789 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971794 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971799 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971812 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971819 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971825 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971833 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971839 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.972429 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.972753 4824 server.go:1280] "Started kubelet" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.973390 4824 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.973395 4824 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 21 11:10:57 crc systemd[1]: Started Kubernetes Kubelet. Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.973808 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.976306 4824 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978242 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978288 4824 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.978732 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978777 4824 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978879 4824 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978777 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 23:23:38.542282267 +0000 UTC Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978794 4824 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.979765 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="200ms" Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.980001 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.980151 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980504 4824 server.go:460] "Adding debug handlers to kubelet server" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980557 4824 factory.go:55] Registering systemd factory Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980581 4824 factory.go:221] Registration of the systemd container factory successfully Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980881 4824 factory.go:153] Registering CRI-O factory Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980905 4824 factory.go:221] Registration of the crio container factory successfully Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980989 4824 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.981053 4824 factory.go:103] Registering Raw factory Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.981086 4824 manager.go:1196] Started watching for new ooms in manager Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.981630 4824 manager.go:319] Starting recovery of all containers Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.979717 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 192.168.26.120:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188cba8a9250030c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-21 11:10:57.972732684 +0000 UTC m=+0.265761976,LastTimestamp:2026-01-21 11:10:57.972732684 +0000 UTC m=+0.265761976,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.985925 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.985999 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986010 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986024 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986033 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986041 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986049 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986057 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986067 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986076 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986084 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986097 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986105 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986117 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986125 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986132 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986140 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986149 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986157 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986164 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986173 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986247 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986257 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986267 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986275 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986283 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986306 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986318 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986329 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986337 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986346 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986362 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986371 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986380 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986388 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986397 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986405 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986425 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986433 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986441 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986448 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986457 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986466 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986475 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986484 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986492 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986501 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986510 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986519 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986527 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986535 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986543 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986556 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986565 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986574 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986583 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986592 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986599 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986607 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986616 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986626 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986646 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986654 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986661 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986671 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986679 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986688 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986698 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986705 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986714 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986722 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986730 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986737 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986745 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986753 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986761 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986771 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986780 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986787 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986796 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986803 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986811 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986820 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986827 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986836 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986847 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986855 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986863 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986870 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986879 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986886 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986894 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986902 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986910 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986917 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986925 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986932 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986940 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986947 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986981 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986990 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986998 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987006 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987013 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987027 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987036 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987046 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987056 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987070 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987078 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987086 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987096 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987104 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987113 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987121 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987129 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987137 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987146 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987159 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987166 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987175 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987183 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987190 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987198 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987207 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987214 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987222 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987230 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987239 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987248 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987257 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987265 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987272 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987279 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987289 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987296 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987305 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987314 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987323 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987332 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987343 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987351 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987359 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987368 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987376 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987384 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987397 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987406 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987426 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987436 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987446 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989684 4824 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989710 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989722 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989731 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989739 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989748 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989757 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989765 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989773 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989781 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989789 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989798 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989806 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989813 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989822 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989835 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989851 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989859 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989867 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989877 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989886 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989904 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989916 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989923 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989934 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989941 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989971 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989981 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989989 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989998 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990006 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990018 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990027 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990035 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990043 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990051 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990061 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990070 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990087 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990096 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990104 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990115 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990125 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990133 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990142 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990149 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990163 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990173 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990184 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990196 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990204 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990212 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990224 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990237 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990245 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990253 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990261 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990273 4824 reconstruct.go:97] "Volume reconstruction finished" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990280 4824 reconciler.go:26] "Reconciler: start to sync state" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.997481 4824 manager.go:324] Recovery completed Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.007670 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.009201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.009284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.009346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.010734 4824 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.010797 4824 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.010852 4824 state_mem.go:36] "Initialized new in-memory state store" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.041482 4824 policy_none.go:49] "None policy: Start" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.042109 4824 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.042135 4824 state_mem.go:35] "Initializing new in-memory state store" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.046489 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.047706 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.048054 4824 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.048088 4824 kubelet.go:2335] "Starting kubelet main sync loop" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.048131 4824 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.048539 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.048576 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.079177 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.080597 4824 manager.go:334] "Starting Device Plugin manager" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.080633 4824 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.080645 4824 server.go:79] "Starting device plugin registration server" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.080969 4824 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.080990 4824 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.081147 4824 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.081216 4824 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.081227 4824 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.085870 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.148749 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.148826 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.149634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.149671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.149681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.149811 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150092 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150985 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.151070 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.151361 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.151836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.151883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.151893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152032 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152127 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152179 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153046 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153203 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153203 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153227 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153889 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153908 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.154495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.154521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.154530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.180397 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="400ms" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.181412 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.182076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.182100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.182110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.182125 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.182352 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192628 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192672 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192685 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192699 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192801 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192911 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192946 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192987 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.193016 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.193035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.193050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.293986 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294124 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294155 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294234 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294241 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294294 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294357 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294362 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294248 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294339 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294445 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294460 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.382707 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.383986 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.384097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.384186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.384296 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.384734 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.493041 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.499385 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.516108 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.516734 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-edf769613ff88e9ab714047d66dcc4a41f0950c889ae2f5fe088499d67ab3ed3 WatchSource:0}: Error finding container edf769613ff88e9ab714047d66dcc4a41f0950c889ae2f5fe088499d67ab3ed3: Status 404 returned error can't find the container with id edf769613ff88e9ab714047d66dcc4a41f0950c889ae2f5fe088499d67ab3ed3 Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.518298 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a4ddd39b572f16102ce0dd8128f64021ff327707e7b38e42cd31f31d0ebdd22d WatchSource:0}: Error finding container a4ddd39b572f16102ce0dd8128f64021ff327707e7b38e42cd31f31d0ebdd22d: Status 404 returned error can't find the container with id a4ddd39b572f16102ce0dd8128f64021ff327707e7b38e42cd31f31d0ebdd22d Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.525441 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4ad861050920fe34134874ba9dc73c919058a01b36434d0a996bb3f687549498 WatchSource:0}: Error finding container 4ad861050920fe34134874ba9dc73c919058a01b36434d0a996bb3f687549498: Status 404 returned error can't find the container with id 4ad861050920fe34134874ba9dc73c919058a01b36434d0a996bb3f687549498 Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.530014 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.534872 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.546524 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-74643405e4bd126e4c118bb553357012409644db3d7dc245d90ad17473d46cfa WatchSource:0}: Error finding container 74643405e4bd126e4c118bb553357012409644db3d7dc245d90ad17473d46cfa: Status 404 returned error can't find the container with id 74643405e4bd126e4c118bb553357012409644db3d7dc245d90ad17473d46cfa Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.547144 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b7b5a9d4ae92b18983fc96eac50835eb2262c50f8125af46811cfbd1c64137a4 WatchSource:0}: Error finding container b7b5a9d4ae92b18983fc96eac50835eb2262c50f8125af46811cfbd1c64137a4: Status 404 returned error can't find the container with id b7b5a9d4ae92b18983fc96eac50835eb2262c50f8125af46811cfbd1c64137a4 Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.581443 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="800ms" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.784857 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.785812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.785851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.785860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.785881 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.786320 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.883394 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.883644 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.974482 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.979802 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 01:24:41.62253628 +0000 UTC Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.051459 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc" exitCode=0 Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.051521 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.051595 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"edf769613ff88e9ab714047d66dcc4a41f0950c889ae2f5fe088499d67ab3ed3"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.051691 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.052544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.052570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.052579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.053618 4824 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875" exitCode=0 Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.053718 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.053750 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a4ddd39b572f16102ce0dd8128f64021ff327707e7b38e42cd31f31d0ebdd22d"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.053838 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.054446 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.056473 4824 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422" exitCode=0 Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.056537 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.056567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b7b5a9d4ae92b18983fc96eac50835eb2262c50f8125af46811cfbd1c64137a4"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.056675 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.057328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.057353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.057363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.057944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.058014 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"74643405e4bd126e4c118bb553357012409644db3d7dc245d90ad17473d46cfa"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.059294 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6" exitCode=0 Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.059331 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.059353 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4ad861050920fe34134874ba9dc73c919058a01b36434d0a996bb3f687549498"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.059437 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.060068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.060094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.060103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: W0121 11:10:59.167129 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:59 crc kubenswrapper[4824]: E0121 11:10:59.167210 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:59 crc kubenswrapper[4824]: W0121 11:10:59.188181 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:59 crc kubenswrapper[4824]: E0121 11:10:59.188240 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:59 crc kubenswrapper[4824]: E0121 11:10:59.382182 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="1.6s" Jan 21 11:10:59 crc kubenswrapper[4824]: W0121 11:10:59.444526 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:59 crc kubenswrapper[4824]: E0121 11:10:59.444584 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.586427 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.588783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.588815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.588825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.588844 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:10:59 crc kubenswrapper[4824]: E0121 11:10:59.589294 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.980517 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 20:56:54.433665752 +0000 UTC Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.062844 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1fd58f2438b6f9cbb2c37687e39f02cdf96a029477b1ae33f6dc9f3b70ecc926"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.062937 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.064192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.064216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.064224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.065510 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.065532 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.065542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.065593 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.066175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.066192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.066199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.068149 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.068185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.068195 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.068264 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.069071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.069093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.069100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.070395 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb" exitCode=0 Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.070444 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.070527 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.071000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.071026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.071035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072898 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072999 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.073478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.073512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.073521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.099155 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.860107 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.865169 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.981203 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 06:18:45.871212848 +0000 UTC Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.071422 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.076445 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48" exitCode=0 Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.076541 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.076989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48"} Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077029 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077093 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077112 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.190283 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.190876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.190903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.190913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.190928 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.981429 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 06:01:05.041617624 +0000 UTC Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083240 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29"} Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083305 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9"} Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb"} Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083333 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681"} Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107"} Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083317 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083409 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083441 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083349 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.981900 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 19:34:53.295504302 +0000 UTC Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.695045 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.695341 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.696248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.696272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.696281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.714631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.714867 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.715531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.715560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.715569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.896826 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.981981 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 02:41:02.681865632 +0000 UTC Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.086004 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.086769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.086806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.086815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.879469 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.879600 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.880519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.880556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.880565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.982539 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 05:22:46.275144861 +0000 UTC Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.656444 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.656558 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.657378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.657400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.657409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.788621 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.788740 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.789486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.789506 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.789513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.983156 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 15:24:02.024830589 +0000 UTC Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.054634 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.090161 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.092121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.092165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.092176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.897224 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.897635 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.984098 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 01:14:23.252791731 +0000 UTC Jan 21 11:11:07 crc kubenswrapper[4824]: I0121 11:11:07.984837 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 11:31:12.854586727 +0000 UTC Jan 21 11:11:08 crc kubenswrapper[4824]: E0121 11:11:08.085993 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 21 11:11:08 crc kubenswrapper[4824]: I0121 11:11:08.985429 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 16:11:29.852049635 +0000 UTC Jan 21 11:11:09 crc kubenswrapper[4824]: I0121 11:11:09.587345 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 21 11:11:09 crc kubenswrapper[4824]: I0121 11:11:09.587409 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 21 11:11:09 crc kubenswrapper[4824]: I0121 11:11:09.600217 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 21 11:11:09 crc kubenswrapper[4824]: I0121 11:11:09.600262 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 21 11:11:09 crc kubenswrapper[4824]: I0121 11:11:09.986423 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 17:06:48.926253659 +0000 UTC Jan 21 11:11:10 crc kubenswrapper[4824]: I0121 11:11:10.020447 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 21 11:11:10 crc kubenswrapper[4824]: I0121 11:11:10.020505 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 21 11:11:10 crc kubenswrapper[4824]: I0121 11:11:10.083189 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 21 11:11:10 crc kubenswrapper[4824]: I0121 11:11:10.083235 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 21 11:11:10 crc kubenswrapper[4824]: I0121 11:11:10.986687 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 08:56:00.015504719 +0000 UTC Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.075517 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.075627 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.075984 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.076047 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.076373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.076399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.076408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.078765 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098364 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098644 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098678 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.987773 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 12:31:04.691331515 +0000 UTC Jan 21 11:11:12 crc kubenswrapper[4824]: I0121 11:11:12.988491 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 17:33:37.525259014 +0000 UTC Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.698621 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.698733 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.699558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.699583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.699592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.989017 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 14:45:26.821735303 +0000 UTC Jan 21 11:11:14 crc kubenswrapper[4824]: E0121 11:11:14.570662 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.572042 4824 trace.go:236] Trace[1618589615]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 11:11:01.321) (total time: 13250ms): Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[1618589615]: ---"Objects listed" error: 13249ms (11:11:14.571) Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[1618589615]: [13.250012041s] [13.250012041s] END Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.572063 4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.572437 4824 trace.go:236] Trace[479225197]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 11:11:01.245) (total time: 13327ms): Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[479225197]: ---"Objects listed" error: 13326ms (11:11:14.572) Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[479225197]: [13.327010649s] [13.327010649s] END Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.572460 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 21 11:11:14 crc kubenswrapper[4824]: E0121 11:11:14.573032 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.573349 4824 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.573575 4824 trace.go:236] Trace[577277594]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 11:11:01.344) (total time: 13229ms): Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[577277594]: ---"Objects listed" error: 13229ms (11:11:14.573) Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[577277594]: [13.229062201s] [13.229062201s] END Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.573594 4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.574207 4824 trace.go:236] Trace[1885941545]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 11:11:01.272) (total time: 13301ms): Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[1885941545]: ---"Objects listed" error: 13301ms (11:11:14.574) Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[1885941545]: [13.301572103s] [13.301572103s] END Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.574228 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.606172 4824 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.611110 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.616058 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.621616 4824 csr.go:261] certificate signing request csr-pz9z5 is approved, waiting to be issued Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.626810 4824 csr.go:257] certificate signing request csr-pz9z5 is issued Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.975684 4824 apiserver.go:52] "Watching apiserver" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.977825 4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978142 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978467 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978482 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978592 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978689 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978804 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:14 crc kubenswrapper[4824]: E0121 11:11:14.978856 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.979070 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:14 crc kubenswrapper[4824]: E0121 11:11:14.979135 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:14 crc kubenswrapper[4824]: E0121 11:11:14.979068 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.979831 4824 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.980269 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.980506 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.980517 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.980590 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.980604 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.981123 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.981647 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.981939 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.982107 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.989087 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 18:51:07.535069084 +0000 UTC Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.991994 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.000489 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.013572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.022415 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.029137 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.036791 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.043403 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.049719 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.056243 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.075935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076000 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076022 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076051 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076066 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076125 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076139 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076155 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076170 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076185 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076202 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076230 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076245 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076259 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076273 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076270 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076291 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076372 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076404 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076433 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076460 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076475 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076491 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076507 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076544 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076557 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076559 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076578 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076596 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076610 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076626 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076639 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076653 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076680 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076708 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076721 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076726 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076727 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076792 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076830 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076866 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076917 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076931 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076992 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077010 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077014 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077040 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077056 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077103 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077117 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077131 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077148 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077164 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077178 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077206 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077219 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077269 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077299 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077319 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077336 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077366 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077375 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077380 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077381 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077442 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077468 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077477 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077488 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077542 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077570 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077576 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077591 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077610 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077627 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077645 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077651 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077657 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077663 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077733 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077745 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077749 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077788 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077787 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077805 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077817 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077835 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077867 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077882 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077910 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077916 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077941 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077975 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078006 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078018 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078023 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078037 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078055 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078074 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078107 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078139 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078154 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078161 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078169 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078186 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078201 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078232 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078247 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078281 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078296 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078329 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078362 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078392 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078408 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078423 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078439 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078454 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078470 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078486 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078501 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078517 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078533 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078548 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078563 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078577 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078594 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078610 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078625 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078640 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078657 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078672 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078688 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078702 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078718 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078733 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078748 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078763 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078781 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078799 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078816 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078832 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078847 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078864 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078879 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078919 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078934 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078950 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079009 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079040 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079090 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079108 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079140 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079156 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079172 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079190 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079207 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079223 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079238 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079252 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079271 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079286 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079320 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079335 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079351 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079367 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079383 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079399 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079430 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079462 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079478 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079494 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079510 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079526 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079542 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079558 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079688 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079710 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079724 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079742 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079757 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079776 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079811 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079878 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079889 4824 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079899 4824 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079909 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079919 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079928 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080255 4824 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080270 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080316 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078273 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078305 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078340 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078389 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078471 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078493 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078525 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078542 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078613 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078626 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078646 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078670 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078743 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078765 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078797 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078876 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078948 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079124 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079226 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079264 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079330 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079407 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079455 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079500 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079510 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079668 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079701 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079866 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080063 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080058 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080070 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080171 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080300 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080321 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080355 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080512 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080623 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080736 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.081740 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.081758 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.081864 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.082020 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:15.581952136 +0000 UTC m=+17.874981427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.083206 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.083262 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.083851 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084122 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084139 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084206 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084198 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084240 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084383 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084419 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084602 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084725 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084871 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084887 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085172 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085247 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085271 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085283 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085298 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085188 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085454 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085476 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085580 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085623 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085661 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085680 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085696 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086085 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086234 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086292 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086331 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086420 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086849 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086944 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086950 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.087891 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088164 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088292 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088318 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088647 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088782 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088851 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089013 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089113 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089177 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089143 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089282 4824 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089372 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089465 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089505 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089642 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089754 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.089834 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.089884 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:15.589869401 +0000 UTC m=+17.882898693 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089993 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089998 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.090094 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.090126 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:15.59011691 +0000 UTC m=+17.883146203 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090133 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090160 4824 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090268 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090314 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090337 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089834 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090637 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090674 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090687 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090697 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090726 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090737 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090746 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091086 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091107 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091120 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091135 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091144 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091209 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091291 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091334 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091541 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.092245 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.092287 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.092381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093311 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093718 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093906 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093912 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.094073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.094102 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.095508 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.095973 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.097082 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-lv4rn"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.097120 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.097358 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.098085 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.099814 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.100024 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.100255 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.100835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.101918 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.102977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.105200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.105428 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.105689 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.105827 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.106040 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.106059 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.106073 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.106084 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.106127 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.106133 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:15.606122217 +0000 UTC m=+17.899151509 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.106215 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.106259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.106279 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.107423 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.108062 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.108151 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.108179 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.108255 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:15.608236301 +0000 UTC m=+17.901265593 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.108262 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.108451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.108629 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.108540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109132 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109189 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109220 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109280 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109413 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109636 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109430 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109559 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109882 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109654 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109951 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.110040 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.110059 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.110101 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.110993 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.111235 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.111300 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.111511 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.111810 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.112368 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.114463 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.116847 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569" exitCode=255 Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.117539 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569"} Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.119595 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.119791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.120801 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.120836 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.123192 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.124145 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.125274 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.125732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.127897 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.129996 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.133594 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.136213 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.142126 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.142320 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.142536 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.142815 4824 scope.go:117] "RemoveContainer" containerID="af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.147562 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.153267 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.159420 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.165451 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.174398 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.180378 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.186636 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.191941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192152 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c1267539-9adf-466f-8027-8ede009d19a0-hosts-file\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbmfl\" (UniqueName: \"kubernetes.io/projected/c1267539-9adf-466f-8027-8ede009d19a0-kube-api-access-dbmfl\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192276 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192293 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192303 4824 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192312 4824 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192320 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192329 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192337 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192344 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192355 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192363 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192371 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192379 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192386 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192394 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192402 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192410 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192418 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192426 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192434 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192442 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192451 4824 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192459 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192467 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192475 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192482 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192491 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192498 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192505 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192513 4824 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192520 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192528 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192534 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192542 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192549 4824 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192557 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192566 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192574 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192582 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192590 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192598 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192606 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192614 4824 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192622 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192630 4824 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192638 4824 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192646 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192655 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192663 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192671 4824 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192678 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192686 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192692 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192699 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192707 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192718 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192727 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192735 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192743 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192750 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192759 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192766 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192779 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192787 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192794 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192802 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192811 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192818 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192825 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192833 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192841 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192848 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192856 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192865 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192872 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192880 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192887 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192894 4824 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192902 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192990 4824 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193002 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193009 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193018 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193027 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193035 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193042 4824 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193051 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193060 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193069 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193077 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193085 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193092 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193099 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193107 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193115 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193122 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193130 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193137 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193151 4824 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193158 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193166 4824 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193174 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193181 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193189 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193197 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193283 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193294 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193302 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193312 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193319 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193326 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193334 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193330 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193344 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193448 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193462 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193472 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193481 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193491 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193500 4824 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193509 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193518 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193525 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193535 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193544 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193551 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193559 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193567 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193575 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193583 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193591 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193599 4824 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193606 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193614 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193622 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193630 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193637 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193645 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193653 4824 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193660 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193668 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193675 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193682 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193692 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193701 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193711 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193720 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193727 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193735 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193743 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193753 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193761 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193769 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193777 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193785 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193792 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193801 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193809 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193817 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193801 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193824 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193932 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193943 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193952 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193977 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193999 4824 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194007 4824 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194016 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194023 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194031 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194039 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194047 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194054 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194062 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194071 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.199562 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.207541 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.212717 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.225283 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.234910 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.289659 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.294512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c1267539-9adf-466f-8027-8ede009d19a0-hosts-file\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.294558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbmfl\" (UniqueName: \"kubernetes.io/projected/c1267539-9adf-466f-8027-8ede009d19a0-kube-api-access-dbmfl\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.294628 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c1267539-9adf-466f-8027-8ede009d19a0-hosts-file\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.296467 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.299017 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-f6a34848b7ccc79f1145a419de2da50011e2e9a8aecedf718fec38ac701b9547 WatchSource:0}: Error finding container f6a34848b7ccc79f1145a419de2da50011e2e9a8aecedf718fec38ac701b9547: Status 404 returned error can't find the container with id f6a34848b7ccc79f1145a419de2da50011e2e9a8aecedf718fec38ac701b9547 Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.302243 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.307352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbmfl\" (UniqueName: \"kubernetes.io/projected/c1267539-9adf-466f-8027-8ede009d19a0-kube-api-access-dbmfl\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.310081 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-5ae7529fc08fbc1bc45653348ce3eaf2bb6e0c95ad21544f7ef8f6655227058d WatchSource:0}: Error finding container 5ae7529fc08fbc1bc45653348ce3eaf2bb6e0c95ad21544f7ef8f6655227058d: Status 404 returned error can't find the container with id 5ae7529fc08fbc1bc45653348ce3eaf2bb6e0c95ad21544f7ef8f6655227058d Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.425674 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.442658 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1267539_9adf_466f_8027_8ede009d19a0.slice/crio-44a63cef2a9562f6b836d0eea1fdc74e6cf08c709d90570463cdcccabafa476c WatchSource:0}: Error finding container 44a63cef2a9562f6b836d0eea1fdc74e6cf08c709d90570463cdcccabafa476c: Status 404 returned error can't find the container with id 44a63cef2a9562f6b836d0eea1fdc74e6cf08c709d90570463cdcccabafa476c Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.452278 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-zqjg9"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.452521 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-x94nb"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.452670 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.452899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.456574 4824 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.456620 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.456625 4824 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.456651 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.456750 4824 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.456772 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457072 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457155 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457356 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-gjtnh"] Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.457383 4824 reflector.go:561] object-"openshift-multus"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.457412 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.457453 4824 reflector.go:561] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": failed to list *v1.Secret: secrets "machine-config-daemon-dockercfg-r5tcq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.457466 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-r5tcq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-config-daemon-dockercfg-r5tcq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457505 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457525 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457531 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457771 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.458093 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-htkvk"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.458727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462102 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462239 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462342 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462506 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462602 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462701 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462813 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.463126 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.464889 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.467562 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.475464 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.481269 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.488415 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-bin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495110 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-etc-kubernetes\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495127 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495118 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495152 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cnibin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-conf-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495197 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495211 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-netns\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/33f3d922-4ffe-409b-a49a-d88c85898260-proxy-tls\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpwst\" (UniqueName: \"kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495359 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-multus-certs\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495386 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495424 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-daemon-config\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495438 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/33f3d922-4ffe-409b-a49a-d88c85898260-rootfs\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgr8g\" (UniqueName: \"kubernetes.io/projected/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-kube-api-access-xgr8g\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495490 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495521 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-socket-dir-parent\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495534 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495546 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495572 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495586 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495601 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495614 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-system-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495640 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-os-release\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-hostroot\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495686 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495715 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cnibin\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495732 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495757 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-k8s-cni-cncf-io\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcf8q\" (UniqueName: \"kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495879 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q78mk\" (UniqueName: \"kubernetes.io/projected/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-kube-api-access-q78mk\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33f3d922-4ffe-409b-a49a-d88c85898260-mcd-auth-proxy-config\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496113 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-multus\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496132 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-kubelet\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496173 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-os-release\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496191 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496219 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496313 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-system-cni-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.502369 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.510051 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.516736 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.523188 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.528504 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.535670 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.541590 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.548593 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.559271 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.566867 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.573843 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.580418 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.586352 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.592796 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597751 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgr8g\" (UniqueName: \"kubernetes.io/projected/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-kube-api-access-xgr8g\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597771 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.597784 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:16.597768627 +0000 UTC m=+18.890797919 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597840 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-socket-dir-parent\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597883 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-os-release\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-os-release\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597977 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-hostroot\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598011 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-socket-dir-parent\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598034 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-hostroot\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598065 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-system-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598105 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598133 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-system-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598137 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-k8s-cni-cncf-io\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-k8s-cni-cncf-io\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598158 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598195 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598293 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598300 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcf8q\" (UniqueName: \"kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598321 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cnibin\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598339 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-multus\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598369 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-kubelet\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-multus\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598402 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q78mk\" (UniqueName: \"kubernetes.io/projected/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-kube-api-access-q78mk\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33f3d922-4ffe-409b-a49a-d88c85898260-mcd-auth-proxy-config\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598426 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-kubelet\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598440 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-os-release\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598540 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598552 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-os-release\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.598587 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598425 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cnibin\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.598599 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.598635 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:16.598623977 +0000 UTC m=+18.891653269 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-system-cni-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.598667 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:16.598653643 +0000 UTC m=+18.891682934 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598678 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-system-cni-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598681 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-etc-kubernetes\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598717 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-etc-kubernetes\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598724 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598740 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-bin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598757 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-bin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598788 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cnibin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598802 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598826 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-conf-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cnibin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598843 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598897 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-netns\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598876 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-conf-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598913 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-multus-certs\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33f3d922-4ffe-409b-a49a-d88c85898260-mcd-auth-proxy-config\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598931 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-multus-certs\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598919 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-netns\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598940 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599005 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599039 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/33f3d922-4ffe-409b-a49a-d88c85898260-proxy-tls\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpwst\" (UniqueName: \"kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-daemon-config\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599132 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/33f3d922-4ffe-409b-a49a-d88c85898260-rootfs\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599184 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/33f3d922-4ffe-409b-a49a-d88c85898260-rootfs\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599378 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599453 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599895 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-daemon-config\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.601112 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.601976 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/33f3d922-4ffe-409b-a49a-d88c85898260-proxy-tls\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.621564 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.627701 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-21 11:06:14 +0000 UTC, rotation deadline is 2026-10-25 14:23:32.222228902 +0000 UTC Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.627745 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6651h12m16.594486s for next certificate rotation Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.689191 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcf8q\" (UniqueName: \"kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.699500 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.699575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699657 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699677 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699687 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699720 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:16.699709197 +0000 UTC m=+18.992738489 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699657 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699768 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699778 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699812 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:16.699803837 +0000 UTC m=+18.992833128 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.740782 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.783909 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:15Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.805773 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.815008 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.823721 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:15Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.844567 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.868818 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.887106 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:15Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.896977 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26b1108a_4b98_4e39_a3cd_e0d055089fd8.slice/crio-b233a1680204a0985feed02a396e93f06104160dd5b2e0d82128d39c672a1129 WatchSource:0}: Error finding container b233a1680204a0985feed02a396e93f06104160dd5b2e0d82128d39c672a1129: Status 404 returned error can't find the container with id b233a1680204a0985feed02a396e93f06104160dd5b2e0d82128d39c672a1129 Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.922793 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:15Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.963418 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:15Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.989995 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 03:25:33.709898958 +0000 UTC Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.006157 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.043506 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.049047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.049140 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.051906 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.052539 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.053558 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.054155 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.055048 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.055518 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.056033 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.056916 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.057501 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.058345 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.058767 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.059689 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.060139 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.060700 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.061652 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.062388 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.063228 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.063555 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.064342 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.065446 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.065931 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.066793 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.067219 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.068343 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.068710 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.069266 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.070193 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.070655 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.071681 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.072362 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.073132 4824 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.073232 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.074741 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.075556 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.076019 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.077319 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.077866 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.078636 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.079391 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.080462 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.080911 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.082343 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.082574 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.083504 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.084842 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.085688 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.086626 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.087346 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.088531 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.088982 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.089737 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.090175 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.091122 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.091729 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.092348 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.120806 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.120835 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.120846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"661c876c65ddef7d3211b181fa8432b45a9d8715f2d2bed40679e29188b710d9"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.122448 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.122472 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f6a34848b7ccc79f1145a419de2da50011e2e9a8aecedf718fec38ac701b9547"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.123835 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.123986 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.125124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.125254 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.125901 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" exitCode=0 Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.125950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.125986 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"b233a1680204a0985feed02a396e93f06104160dd5b2e0d82128d39c672a1129"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.127194 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5ae7529fc08fbc1bc45653348ce3eaf2bb6e0c95ad21544f7ef8f6655227058d"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.128319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lv4rn" event={"ID":"c1267539-9adf-466f-8027-8ede009d19a0","Type":"ContainerStarted","Data":"eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.128347 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lv4rn" event={"ID":"c1267539-9adf-466f-8027-8ede009d19a0","Type":"ContainerStarted","Data":"44a63cef2a9562f6b836d0eea1fdc74e6cf08c709d90570463cdcccabafa476c"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.162840 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.209570 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.242985 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.284843 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.326431 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.365751 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.377733 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.423143 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.466374 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.505255 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.517745 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.529536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgr8g\" (UniqueName: \"kubernetes.io/projected/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-kube-api-access-xgr8g\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.530780 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q78mk\" (UniqueName: \"kubernetes.io/projected/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-kube-api-access-q78mk\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.567644 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.599023 4824 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.599078 4824 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.599083 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy podName:6b9588d1-5876-4c2c-8de5-e5c0ad46f04e nodeName:}" failed. No retries permitted until 2026-01-21 11:11:17.099068855 +0000 UTC m=+19.392098146 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy") pod "multus-x94nb" (UID: "6b9588d1-5876-4c2c-8de5-e5c0ad46f04e") : failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.599158 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy podName:3a8d2cac-2d0f-4894-9bc0-2abd81699c47 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:17.099142583 +0000 UTC m=+19.392171876 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy") pod "multus-additional-cni-plugins-gjtnh" (UID: "3a8d2cac-2d0f-4894-9bc0-2abd81699c47") : failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.605539 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.617111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.617233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.617249 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:18.617235194 +0000 UTC m=+20.910264486 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.617280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.617317 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.617361 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.617369 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:18.617358116 +0000 UTC m=+20.910387409 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.617388 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:18.617381381 +0000 UTC m=+20.910410673 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.643073 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.682543 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.707403 4824 projected.go:288] Couldn't get configMap openshift-machine-config-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.707436 4824 projected.go:194] Error preparing data for projected volume kube-api-access-mpwst for pod openshift-machine-config-operator/machine-config-daemon-zqjg9: failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.707493 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst podName:33f3d922-4ffe-409b-a49a-d88c85898260 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:17.207477426 +0000 UTC m=+19.500506718 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mpwst" (UniqueName: "kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst") pod "machine-config-daemon-zqjg9" (UID: "33f3d922-4ffe-409b-a49a-d88c85898260") : failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.718400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.718470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718571 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718599 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718610 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718646 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:18.71863493 +0000 UTC m=+21.011664221 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718578 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718690 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718701 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718740 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:18.718729519 +0000 UTC m=+21.011758811 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.725693 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.763518 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.777736 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.823708 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.837727 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.886161 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.922951 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.937005 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.983913 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.990970 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 09:10:21.198813534 +0000 UTC Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.023985 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.048908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.048945 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.049022 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.049074 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.063077 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.103465 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.121835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.121873 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.122463 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.122516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.131709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134193 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134252 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134268 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.143429 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.182100 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.222975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpwst\" (UniqueName: \"kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.224168 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.225452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpwst\" (UniqueName: \"kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.264285 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.267346 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x94nb" Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.275255 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b9588d1_5876_4c2c_8de5_e5c0ad46f04e.slice/crio-b30310394ccb704d21dccf9961467707c8f16173677b5f28b9969e359c6eb524 WatchSource:0}: Error finding container b30310394ccb704d21dccf9961467707c8f16173677b5f28b9969e359c6eb524: Status 404 returned error can't find the container with id b30310394ccb704d21dccf9961467707c8f16173677b5f28b9969e359c6eb524 Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.307730 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.327205 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.335835 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33f3d922_4ffe_409b_a49a_d88c85898260.slice/crio-da8510a8b5ef1f15b44d065bf5855b68b621b13a0308b5a3f3e3b09af89ede7f WatchSource:0}: Error finding container da8510a8b5ef1f15b44d065bf5855b68b621b13a0308b5a3f3e3b09af89ede7f: Status 404 returned error can't find the container with id da8510a8b5ef1f15b44d065bf5855b68b621b13a0308b5a3f3e3b09af89ede7f Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.343819 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.358134 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.369154 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a8d2cac_2d0f_4894_9bc0_2abd81699c47.slice/crio-e52ebed3ade7f1249cea45e214ee412ca574617c3d80124ed44d587e2650f1d9 WatchSource:0}: Error finding container e52ebed3ade7f1249cea45e214ee412ca574617c3d80124ed44d587e2650f1d9: Status 404 returned error can't find the container with id e52ebed3ade7f1249cea45e214ee412ca574617c3d80124ed44d587e2650f1d9 Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.385727 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.426052 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.464603 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.503847 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.548994 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.573055 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-jr64s"] Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.573364 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.584674 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.598104 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.617988 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.626305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flrnp\" (UniqueName: \"kubernetes.io/projected/9768c60d-5bee-4286-bced-076e584b04e1-kube-api-access-flrnp\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.626375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9768c60d-5bee-4286-bced-076e584b04e1-host\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.626423 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9768c60d-5bee-4286-bced-076e584b04e1-serviceca\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.637138 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.657224 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.704714 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.727580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9768c60d-5bee-4286-bced-076e584b04e1-host\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.727613 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9768c60d-5bee-4286-bced-076e584b04e1-serviceca\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.727630 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flrnp\" (UniqueName: \"kubernetes.io/projected/9768c60d-5bee-4286-bced-076e584b04e1-kube-api-access-flrnp\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.727681 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9768c60d-5bee-4286-bced-076e584b04e1-host\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.728511 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9768c60d-5bee-4286-bced-076e584b04e1-serviceca\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.748262 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.769630 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flrnp\" (UniqueName: \"kubernetes.io/projected/9768c60d-5bee-4286-bced-076e584b04e1-kube-api-access-flrnp\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.773258 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.774703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.774725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.774733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.774796 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.828643 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.838384 4824 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.838606 4824 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.839449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.839476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.839484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.839495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.839504 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.847769 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.850255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.850287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.850297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.850310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.850317 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.858475 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.860882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.860972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.860985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.860997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.861019 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.870942 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.873515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.873577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.873598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.873611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.873619 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.882135 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.882398 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885168 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885811 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885868 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.892613 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9768c60d_5bee_4286_bced_076e584b04e1.slice/crio-b14e560c96f826d1223c405141516af2082c14ca9109faa9a67b058b78fd0c1a WatchSource:0}: Error finding container b14e560c96f826d1223c405141516af2082c14ca9109faa9a67b058b78fd0c1a: Status 404 returned error can't find the container with id b14e560c96f826d1223c405141516af2082c14ca9109faa9a67b058b78fd0c1a Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.896813 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.896914 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.897809 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.897888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.898053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.898198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.898332 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.909070 4824 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.909926 4824 reflector.go:484] object-"openshift-image-registry"/"node-ca-dockercfg-4777p": watch of *v1.Secret ended with: very short watch: object-"openshift-image-registry"/"node-ca-dockercfg-4777p": Unexpected watch close - watch lasted less than a second and no items received Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.910365 4824 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.913426 4824 reflector.go:484] object-"openshift-image-registry"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.913509 4824 reflector.go:484] object-"openshift-image-registry"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.913574 4824 reflector.go:484] object-"openshift-image-registry"/"image-registry-certificates": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"image-registry-certificates": Unexpected watch close - watch lasted less than a second and no items received Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.931688 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.963221 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.991400 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 19:25:27.25383742 +0000 UTC Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.000202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.000233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.000242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.000256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.000264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.003852 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.043334 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.048645 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.048757 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.082426 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.101864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.101887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.101895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.101905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.101914 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.122362 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.138724 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba" exitCode=0 Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.138811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.138848 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerStarted","Data":"e52ebed3ade7f1249cea45e214ee412ca574617c3d80124ed44d587e2650f1d9"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.139942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jr64s" event={"ID":"9768c60d-5bee-4286-bced-076e584b04e1","Type":"ContainerStarted","Data":"b14e560c96f826d1223c405141516af2082c14ca9109faa9a67b058b78fd0c1a"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.141783 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.141807 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.141820 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"da8510a8b5ef1f15b44d065bf5855b68b621b13a0308b5a3f3e3b09af89ede7f"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.147255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerStarted","Data":"75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.147296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerStarted","Data":"b30310394ccb704d21dccf9961467707c8f16173677b5f28b9969e359c6eb524"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.164930 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.203984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.204027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.204037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.204052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.204064 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.207646 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.244509 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.283264 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.309092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.309377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.309387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.309401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.309410 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.324709 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.366317 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.404226 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.411533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.411564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.411574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.411588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.411597 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.448275 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.484550 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.513483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.513514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.513523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.513536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.513544 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.524203 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.565175 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.605909 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.615469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.615497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.615506 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.615520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.615534 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.634127 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.634219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.634242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.634271 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:22.634254482 +0000 UTC m=+24.927283774 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.634301 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.634336 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:22.634326488 +0000 UTC m=+24.927355779 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.634426 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.634515 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:22.634484828 +0000 UTC m=+24.927514130 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.643065 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.685874 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.717746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.717785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.717794 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.717808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.717816 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.723569 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.735066 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.735111 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735172 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735198 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735210 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735219 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735239 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735249 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735260 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:22.73524411 +0000 UTC m=+25.028273412 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735285 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:22.735274258 +0000 UTC m=+25.028303549 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.764047 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.807515 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.819421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.819461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.819471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.819485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.819494 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.843123 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.877725 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.906219 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.921322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.921349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.921359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.921373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.921382 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.943913 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.983129 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.992322 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 05:27:30.994738474 +0000 UTC Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.022900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.022925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.022935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.022947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.022968 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.023810 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.048747 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:19 crc kubenswrapper[4824]: E0121 11:11:19.048840 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.048751 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:19 crc kubenswrapper[4824]: E0121 11:11:19.048943 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.063665 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.077796 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.122624 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.125005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.125049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.125059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.125071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.125079 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.152639 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.153931 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816" exitCode=0 Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.154005 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.155004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jr64s" event={"ID":"9768c60d-5bee-4286-bced-076e584b04e1","Type":"ContainerStarted","Data":"91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.165169 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.204264 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.227077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.227114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.227123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.227136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.227171 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.245437 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.283808 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.323598 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.329356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.329383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.329391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.329404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.329412 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.357570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.387081 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.396994 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.431100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.431127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.431136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.431148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.431157 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.447561 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.483419 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.522568 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.533041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.533071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.533080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.533095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.533104 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.562676 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.601975 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.635104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.635141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.635152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.635165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.635174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.647634 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.683534 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.726227 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.737359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.737394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.737404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.737417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.737426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.762817 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.803275 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.839984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.840016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.840039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.840051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.840060 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.842498 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.886532 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.923822 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.942427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.942453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.942462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.942473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.942483 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.964323 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.993341 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 03:32:15.942227212 +0000 UTC Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.007225 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.042408 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.044975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.045010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.045033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.045047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.045056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.048567 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:20 crc kubenswrapper[4824]: E0121 11:11:20.048658 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.084057 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.122851 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.147132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.147158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.147166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.147178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.147187 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.159324 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862" exitCode=0 Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.159405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.163471 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.201922 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.243916 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.249765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.249802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.249811 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.249824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.249832 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.284518 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.324517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.352109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.352136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.352146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.352157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.352166 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.363666 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.402828 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.442265 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.453496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.453527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.453536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.453550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.453559 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.481529 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.529302 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.555595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.555622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.555630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.555644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.555653 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.562120 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.602517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.644574 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.657886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.657926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.657964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.657980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.658008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.683779 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.724968 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.759980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.760015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.760024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.760051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.760059 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.766921 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.805642 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.843531 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.863528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.863583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.863595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.863611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.863621 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.887681 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.897633 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.965926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.965986 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.965997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.966011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.966020 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.994328 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 22:16:12.690104605 +0000 UTC Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.048816 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:21 crc kubenswrapper[4824]: E0121 11:11:21.048901 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.048944 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:21 crc kubenswrapper[4824]: E0121 11:11:21.049064 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.068189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.068220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.068229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.068239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.068247 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.164175 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c" exitCode=0 Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.164365 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.168096 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.168302 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.169232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.169264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.169274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.169286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.169294 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.173226 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.185505 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.185614 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.194557 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.204929 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.212148 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.225458 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.234483 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.244514 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.262509 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.270432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.270459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.270469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.270480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.270489 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.303798 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.342711 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.371713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.371737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.371748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.371765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.371774 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.384402 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.423171 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.462735 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.473909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.473931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.473940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.474009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.474022 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.507370 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.546922 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.576150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.576172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.576181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.576193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.576202 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.583505 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.623609 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.663913 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.678523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.678560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.678569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.678582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.678590 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.703252 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.742138 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.779915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.779940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.779948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.779974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.779982 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.784142 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.822821 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.863025 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.881891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.881923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.881932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.881944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.881952 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.905659 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.940663 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983456 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983688 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.994698 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 18:50:49.442414489 +0000 UTC Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.023508 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.048610 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.048695 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.062173 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.084928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.084982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.084993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.085006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.085015 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.102374 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.173584 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758" exitCode=0 Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.173636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.173702 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.174031 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.186997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.187019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.187027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.187039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.187061 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.191672 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.194841 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.199930 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.248374 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.265498 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.289115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.289156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.289170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.289186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.289196 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.305319 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.343535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.383166 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.391494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.391522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.391530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.391542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.391552 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.426979 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.463415 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.493985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.494014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.494022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.494036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.494056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.503884 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.544553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.583564 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.596225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.596259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.596268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.596281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.596292 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.622231 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.662535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.667776 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.667884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.667906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.667929 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:30.667909826 +0000 UTC m=+32.960939119 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.667982 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.668026 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:30.668014034 +0000 UTC m=+32.961043326 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.668092 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.668143 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:30.668130364 +0000 UTC m=+32.961159656 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.698040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.698090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.698100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.698113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.698124 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.704712 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.743548 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.768695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.768769 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768870 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768874 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768920 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768934 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768890 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768992 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.769003 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:30.76898634 +0000 UTC m=+33.062015642 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.769032 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:30.769019003 +0000 UTC m=+33.062048295 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.781399 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.799596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.799624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.799632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.799644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.799651 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.822081 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.868305 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.900922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.900943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.900953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.900984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.900992 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.904417 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.946517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.982425 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.995542 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 09:40:14.721699282 +0000 UTC Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.003094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.003127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.003136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.003149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.003156 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.022100 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.049029 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.049131 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:23 crc kubenswrapper[4824]: E0121 11:11:23.049249 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:23 crc kubenswrapper[4824]: E0121 11:11:23.049401 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.064453 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.103407 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.104597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.104626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.104635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.104646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.104654 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.143864 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.177488 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/0.log" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.179555 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973" exitCode=1 Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.179618 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.180079 4824 scope.go:117] "RemoveContainer" containerID="b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.181889 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630" exitCode=0 Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.181929 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.188232 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.205908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.205942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.205968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.205981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.205990 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.223047 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.263649 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.306859 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.307159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.307185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.307192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.307206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.307214 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.343294 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.384745 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.409191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.409240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.409250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.409264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.409273 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.428799 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"message\\\":\\\"ices_controller.go:444] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643711 6069 services_controller.go:445] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643033 6069 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-jr64s\\\\nI0121 11:11:22.642992 6069 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0121 11:11:22.643725 6069 services_controller.go:451] Built service openshift-operator-lifecycle-manager/olm-operator-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.462465 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.504136 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.511607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.511644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.511654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.511668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.511683 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.543310 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.583687 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.614347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.614391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.614406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.614421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.614430 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.623509 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.669085 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.703818 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.716356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.716390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.716400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.716414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.716422 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.744770 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.782858 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.818683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.818715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.818725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.818738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.818746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.823553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.863514 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.904912 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.920412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.920454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.920463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.920478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.920486 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.996101 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 05:26:17.249976943 +0000 UTC Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.022868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.022911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.022920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.022934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.022941 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.049310 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:24 crc kubenswrapper[4824]: E0121 11:11:24.049437 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.124550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.124582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.124590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.124602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.124610 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.185109 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/1.log" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.185543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/0.log" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.187709 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" exitCode=1 Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.187768 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.187806 4824 scope.go:117] "RemoveContainer" containerID="b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.188263 4824 scope.go:117] "RemoveContainer" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" Jan 21 11:11:24 crc kubenswrapper[4824]: E0121 11:11:24.188404 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.190572 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerStarted","Data":"67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.199391 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.208704 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.216419 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.223536 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.225693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.225717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.225725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.225736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.225744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.232920 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.245620 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.253617 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.262660 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.270445 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.304839 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.327532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.327566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.327577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.327592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.327601 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.344611 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.383265 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.427250 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"message\\\":\\\"ices_controller.go:444] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643711 6069 services_controller.go:445] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643033 6069 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-jr64s\\\\nI0121 11:11:22.642992 6069 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0121 11:11:22.643725 6069 services_controller.go:451] Built service openshift-operator-lifecycle-manager/olm-operator-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.429942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.429990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.429999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.430015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.430024 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.463098 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.503767 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.531695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.531727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.531736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.531748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.531758 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.542016 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.585043 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.623999 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.633496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.633531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.633543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.633561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.633572 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.665253 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.703207 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.735019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.735047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.735057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.735080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.735089 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.747549 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.783457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.825627 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.836662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.836692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.836703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.836717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.836726 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.866222 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.904202 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.939328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.939359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.939369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.939381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.939390 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.943630 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.985166 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.996308 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 13:03:07.313307472 +0000 UTC Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.022325 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.041309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.041342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.041350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.041364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.041373 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.048589 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:25 crc kubenswrapper[4824]: E0121 11:11:25.048678 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.048588 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:25 crc kubenswrapper[4824]: E0121 11:11:25.048787 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.062738 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.105751 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"message\\\":\\\"ices_controller.go:444] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643711 6069 services_controller.go:445] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643033 6069 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-jr64s\\\\nI0121 11:11:22.642992 6069 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0121 11:11:22.643725 6069 services_controller.go:451] Built service openshift-operator-lifecycle-manager/olm-operator-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.143330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.143420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.143441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.143468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.143483 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.194697 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/1.log" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.198278 4824 scope.go:117] "RemoveContainer" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" Jan 21 11:11:25 crc kubenswrapper[4824]: E0121 11:11:25.198411 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.207826 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.216837 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.223654 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.245334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.245369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.245378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.245390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.245400 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.262241 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.302953 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347124 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347695 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.384235 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.422528 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.449392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.449436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.449447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.449465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.449475 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.463093 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.504364 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.543869 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.551235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.551280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.551290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.551307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.551317 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.582519 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.626050 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.652835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.652866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.652875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.652894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.652903 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.662841 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.702938 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.754417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.754457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.754466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.754479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.754487 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.856269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.856304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.856314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.856326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.856337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.958267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.958306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.958315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.958339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.958352 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.996469 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 18:16:19.8902512 +0000 UTC Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.048861 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:26 crc kubenswrapper[4824]: E0121 11:11:26.049006 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.059568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.059598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.059607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.059620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.059628 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.161760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.161791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.161800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.161815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.161823 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.263106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.263304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.263395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.263454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.263511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.365161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.365187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.365195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.365209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.365217 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.467365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.467387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.467394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.467405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.467412 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.569605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.569641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.569650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.569664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.569678 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.671686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.671720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.671728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.671743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.671753 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.773437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.773465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.773474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.773484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.773492 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.875362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.875391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.875399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.875410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.875419 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.977196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.977357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.977437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.977497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.977555 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.996559 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 16:56:14.98029536 +0000 UTC Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.049220 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:27 crc kubenswrapper[4824]: E0121 11:11:27.049379 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.049223 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:27 crc kubenswrapper[4824]: E0121 11:11:27.049548 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.078778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.078826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.078837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.078849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.078858 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.180859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.180895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.180904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.180917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.180929 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.197916 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8"] Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.198259 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.199481 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.199678 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.214914 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.222770 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.230519 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.238155 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.245517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.251801 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.260702 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.267773 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.278280 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.283057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.283094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.283104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.283116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.283124 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.290385 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.297436 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.305836 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.307124 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9qbj\" (UniqueName: \"kubernetes.io/projected/4698abe4-0f33-4349-a2a4-614de80ae21b-kube-api-access-n9qbj\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.307241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.307335 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.307425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4698abe4-0f33-4349-a2a4-614de80ae21b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.313468 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.321075 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.327618 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.334554 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.384735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.384762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.384772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.384785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.384812 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.408386 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9qbj\" (UniqueName: \"kubernetes.io/projected/4698abe4-0f33-4349-a2a4-614de80ae21b-kube-api-access-n9qbj\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.408434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.408468 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.408482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4698abe4-0f33-4349-a2a4-614de80ae21b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.409037 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.409107 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.412470 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4698abe4-0f33-4349-a2a4-614de80ae21b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.419924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9qbj\" (UniqueName: \"kubernetes.io/projected/4698abe4-0f33-4349-a2a4-614de80ae21b-kube-api-access-n9qbj\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.486569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.486607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.486616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.486631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.486642 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.506947 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: W0121 11:11:27.515491 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4698abe4_0f33_4349_a2a4_614de80ae21b.slice/crio-83061edbe8885031d9d7ad1d9dc528be6ab148f6f36122e2040a9e71cd925887 WatchSource:0}: Error finding container 83061edbe8885031d9d7ad1d9dc528be6ab148f6f36122e2040a9e71cd925887: Status 404 returned error can't find the container with id 83061edbe8885031d9d7ad1d9dc528be6ab148f6f36122e2040a9e71cd925887 Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.588561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.588587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.588595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.588608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.588616 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.690901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.690932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.690941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.690966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.690975 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.793074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.793119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.793130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.793144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.793153 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.895294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.895325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.895334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.895346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.895355 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.996906 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 14:43:16.754203739 +0000 UTC Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.997056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.997078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.997086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.997106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.997116 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.048897 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.049019 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.062302 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.071311 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.076783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.076810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.076823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.076838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.076848 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.081774 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.085032 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.087579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.087607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.087616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.087627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.087634 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.096275 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.097270 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.099822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.099854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.099863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.099875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.099883 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.104952 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.107667 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.109937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.109990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.110002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.110014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.110024 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.115575 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.118327 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.120607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.120630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.120639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.120651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.120660 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.122517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.129647 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.129884 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.130845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.130869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.130878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.130888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.130895 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.131278 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.139470 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.151343 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.158317 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.165924 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.172034 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.179237 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.187442 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.195637 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.206709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" event={"ID":"4698abe4-0f33-4349-a2a4-614de80ae21b","Type":"ContainerStarted","Data":"050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.206746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" event={"ID":"4698abe4-0f33-4349-a2a4-614de80ae21b","Type":"ContainerStarted","Data":"191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.206756 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" event={"ID":"4698abe4-0f33-4349-a2a4-614de80ae21b","Type":"ContainerStarted","Data":"83061edbe8885031d9d7ad1d9dc528be6ab148f6f36122e2040a9e71cd925887"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.215755 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.223401 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.231617 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.232636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.232667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.232678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.232692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.232700 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.238543 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.251256 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.255492 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-9rmnw"] Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.255892 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.255948 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.258931 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.266990 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.278620 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.285570 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.292088 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.298249 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.306736 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.314411 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.316190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.316226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2fgn\" (UniqueName: \"kubernetes.io/projected/8f9bd12b-22ee-44ed-a8b9-c100d043f691-kube-api-access-q2fgn\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.321742 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.334806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.334830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.334838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.334851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.334860 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.335859 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.344054 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.356885 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.371998 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.384572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.397910 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.406616 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.414251 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.417550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.417588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2fgn\" (UniqueName: \"kubernetes.io/projected/8f9bd12b-22ee-44ed-a8b9-c100d043f691-kube-api-access-q2fgn\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.417733 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.417812 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:28.917794601 +0000 UTC m=+31.210823893 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.421430 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.429706 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2fgn\" (UniqueName: \"kubernetes.io/projected/8f9bd12b-22ee-44ed-a8b9-c100d043f691-kube-api-access-q2fgn\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.431463 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.436533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.436676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.436749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.436815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.436871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.441596 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.454863 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.464088 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.471693 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.482196 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.490053 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.499118 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.508389 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.517352 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.538758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.538777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.538785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.538796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.538804 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.640646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.640732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.640810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.640874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.640941 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.731023 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.731665 4824 scope.go:117] "RemoveContainer" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.731852 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.743212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.743244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.743254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.743266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.743275 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.844775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.844808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.844817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.844828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.844836 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.921781 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.921892 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.921937 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:29.921924535 +0000 UTC m=+32.214953827 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.946802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.946836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.946845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.946857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.946865 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.997264 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 02:35:40.001102101 +0000 UTC Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048311 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:29 crc kubenswrapper[4824]: E0121 11:11:29.048330 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:29 crc kubenswrapper[4824]: E0121 11:11:29.048406 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048915 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.151044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.151082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.151090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.151113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.151122 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.253541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.253572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.253581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.253591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.253603 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.355284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.355306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.355314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.355323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.355331 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.456844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.456875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.456886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.456898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.456907 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.558998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.559038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.559047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.559060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.559069 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.661163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.661198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.661207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.661218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.661226 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.763485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.763529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.763539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.763552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.763561 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.865197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.865231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.865239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.865251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.865259 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.930839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:29 crc kubenswrapper[4824]: E0121 11:11:29.930929 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:29 crc kubenswrapper[4824]: E0121 11:11:29.930984 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:31.930973149 +0000 UTC m=+34.224002442 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.967184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.967228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.967238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.967250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.967258 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.997429 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 14:38:13.779757788 +0000 UTC Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.023019 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.032542 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.039985 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.047061 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.048437 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.048480 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.048506 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.048562 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.056079 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.064378 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.068989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.069022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.069033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.069046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.069055 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.076540 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.084305 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.091951 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.101087 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.108020 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.115943 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.127181 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.137243 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.144659 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.152191 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.159990 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.171415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.171445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.171455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.171468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.171476 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.173086 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.273397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.273435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.273443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.273456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.273464 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.375277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.375534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.375604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.375670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.375747 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.478380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.478425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.478435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.478449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.478458 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.580443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.580743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.580753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.580766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.580776 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.683137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.683178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.683190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.683206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.683216 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.739742 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.739835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.739894 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.740001 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.740012 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.740000 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:46.739952343 +0000 UTC m=+49.032981645 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.740070 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:46.740054586 +0000 UTC m=+49.033083879 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.740082 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:46.740076047 +0000 UTC m=+49.033105339 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.785412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.785454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.785463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.785482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.785490 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.841275 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.841332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841455 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841472 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841482 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841487 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841517 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841526 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841531 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:46.841518704 +0000 UTC m=+49.134547996 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841581 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:46.841564361 +0000 UTC m=+49.134593652 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.888217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.888271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.888282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.888297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.888313 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.990443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.990475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.990484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.990502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.990511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.997498 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 06:33:00.023934931 +0000 UTC Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.048434 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:31 crc kubenswrapper[4824]: E0121 11:11:31.048536 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.048435 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:31 crc kubenswrapper[4824]: E0121 11:11:31.048607 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.092917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.092978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.092989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.093000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.093008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.194605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.194635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.194644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.194656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.194664 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.296810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.296839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.296848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.296876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.296885 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.398854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.398898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.398907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.398919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.398927 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.500405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.500449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.500456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.500471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.500481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.602569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.602603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.602611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.602634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.602642 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.703868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.703907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.703916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.703929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.703937 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.805769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.805808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.805816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.805829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.805838 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.907651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.907706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.907717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.907729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.907738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.952615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:31 crc kubenswrapper[4824]: E0121 11:11:31.952752 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:31 crc kubenswrapper[4824]: E0121 11:11:31.952797 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:35.952786356 +0000 UTC m=+38.245815648 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.998227 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 05:01:56.311426 +0000 UTC Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.009630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.009649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.009660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.009671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.009680 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.049205 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.049260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:32 crc kubenswrapper[4824]: E0121 11:11:32.049334 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:32 crc kubenswrapper[4824]: E0121 11:11:32.049408 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.111681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.111708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.111716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.111727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.111735 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.213309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.213332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.213340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.213349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.213356 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.314947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.314996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.315005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.315014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.315021 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.417109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.417151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.417161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.417174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.417183 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.519578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.519618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.519629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.519644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.519652 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.621429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.621461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.621468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.621481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.621490 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.723048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.723121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.723148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.723162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.723173 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.825102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.825145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.825153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.825166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.825175 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.926772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.926804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.926813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.926827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.926837 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.998521 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 11:08:18.52546233 +0000 UTC Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.028985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.029013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.029020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.029030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.029039 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.048397 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.048417 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:33 crc kubenswrapper[4824]: E0121 11:11:33.048477 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:33 crc kubenswrapper[4824]: E0121 11:11:33.048553 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.130341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.130392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.130402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.130416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.130425 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.232228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.232269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.232277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.232290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.232299 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.333608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.333637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.333645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.333657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.333665 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.435453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.435473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.435482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.435491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.435499 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.537505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.537544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.537553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.537569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.537577 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.639336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.639372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.639381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.639394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.639402 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.741300 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.741331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.741340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.741354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.741362 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.843135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.843179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.843188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.843198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.843205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.945225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.945259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.945268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.945281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.945288 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.998819 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 06:43:51.897008128 +0000 UTC Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.047407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.047446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.047457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.047473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.047483 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.048546 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.048562 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:34 crc kubenswrapper[4824]: E0121 11:11:34.048628 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:34 crc kubenswrapper[4824]: E0121 11:11:34.048685 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.149328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.149357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.149365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.149375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.149383 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.251478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.251509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.251518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.251532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.251544 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.353000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.353030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.353040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.353052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.353061 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.455325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.455354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.455363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.455376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.455385 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.556763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.556788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.556797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.556807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.556814 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.659252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.659302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.659312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.659324 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.659333 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.760886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.760911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.760919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.760931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.760940 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.862457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.862480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.862488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.862498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.862505 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.964136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.964178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.964187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.964198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.964206 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.998905 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 21:12:04.790777236 +0000 UTC Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.048488 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:35 crc kubenswrapper[4824]: E0121 11:11:35.048622 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.048492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:35 crc kubenswrapper[4824]: E0121 11:11:35.048740 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.066324 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.066355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.066365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.066379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.066388 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.168314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.168339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.168347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.168357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.168364 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.269751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.269784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.269792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.269804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.269814 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.372060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.372096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.372107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.372121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.372131 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.473577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.473614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.473622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.473637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.473646 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.575531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.575558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.575566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.575576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.575584 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.677914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.677944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.677952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.677975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.677984 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.779771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.779803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.779811 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.779824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.779833 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.881081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.881110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.881119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.881129 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.881137 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.982736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.982775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.982785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.982798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.982815 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.986122 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:35 crc kubenswrapper[4824]: E0121 11:11:35.986225 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:35 crc kubenswrapper[4824]: E0121 11:11:35.986270 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:43.986258054 +0000 UTC m=+46.279287346 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.999493 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 19:38:12.287513634 +0000 UTC Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.048943 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.048949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:36 crc kubenswrapper[4824]: E0121 11:11:36.049158 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:36 crc kubenswrapper[4824]: E0121 11:11:36.049065 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.084704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.084734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.084744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.084754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.084763 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.186142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.186187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.186196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.186208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.186217 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.287539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.287567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.287576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.287586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.287594 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.389074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.389110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.389118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.389132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.389140 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.490990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.491013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.491022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.491033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.491041 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.591988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.592011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.592020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.592032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.592040 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.693345 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.693375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.693384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.693395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.693403 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.795211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.795240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.795249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.795260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.795268 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.896646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.896673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.896680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.896689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.896698 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.998706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.998737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.998744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.998755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.998763 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.999809 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 11:30:17.363330206 +0000 UTC Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.049292 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.049306 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:37 crc kubenswrapper[4824]: E0121 11:11:37.049385 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:37 crc kubenswrapper[4824]: E0121 11:11:37.049466 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.100162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.100202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.100211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.100221 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.100230 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.201641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.201663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.201670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.201680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.201688 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.303669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.303689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.303696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.303705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.303713 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.405553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.405580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.405593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.405603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.405609 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.507813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.507840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.507848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.507857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.507864 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.609436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.609460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.609468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.609477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.609484 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.711479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.711520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.711538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.711552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.711565 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.813276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.813306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.813315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.813326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.813338 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.914696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.914740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.914749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.914768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.914775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.000406 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:34:46.449565449 +0000 UTC Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.016691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.016714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.016722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.016733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.016740 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.049103 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.049197 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.049284 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.049369 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.058676 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.065778 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.073215 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.080224 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.089950 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.097749 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.106718 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.114393 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.118999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.119042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.119053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.119064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.119073 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.125883 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.133469 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.142199 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.150044 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.157656 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.161831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.161854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.161863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.161874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.161882 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.165666 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.169261 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.171486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.171506 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.171515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.171524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.171532 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.172029 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.178596 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.181121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.181149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.181158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.181167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.181174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.184104 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.189086 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191222 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191568 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.199295 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.201630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.201655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.201663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.201672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.201679 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.208951 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.209093 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.221459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.221495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.221507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.221523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.221533 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.323627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.323678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.323688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.323700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.323711 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.426348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.426409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.426419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.426431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.426440 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.528380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.528415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.528423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.528443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.528451 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.630028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.630053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.630063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.630073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.630082 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.732122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.732152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.732162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.732175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.732194 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.834507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.834542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.834551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.834582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.834591 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.936613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.936639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.936647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.936660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.936668 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.001309 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 15:44:27.184701429 +0000 UTC Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.039283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.039333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.039343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.039355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.039363 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.048580 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.048620 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:39 crc kubenswrapper[4824]: E0121 11:11:39.048694 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:39 crc kubenswrapper[4824]: E0121 11:11:39.048778 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.141925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.141979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.141989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.142002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.142011 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.244179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.244227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.244238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.244250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.244259 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.346390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.346429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.346438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.346453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.346463 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.448798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.448836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.448848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.448862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.448871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.551081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.551115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.551125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.551138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.551145 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.653323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.653378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.653387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.653401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.653409 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.755449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.755486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.755494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.755507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.755517 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.857533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.857578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.857587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.857606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.857614 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.959932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.960000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.960016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.960029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.960036 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.001475 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 20:49:52.696614283 +0000 UTC Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.049047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.049088 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:40 crc kubenswrapper[4824]: E0121 11:11:40.049146 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:40 crc kubenswrapper[4824]: E0121 11:11:40.049280 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.050437 4824 scope.go:117] "RemoveContainer" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.061435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.061478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.061489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.061503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.061515 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.164472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.164499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.164508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.164523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.164532 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.235663 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/1.log" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.238439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.238815 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.247993 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.258367 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266419 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266641 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.277205 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.286468 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.301368 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.311837 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.320828 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.335608 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.348845 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.358177 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.369392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.369426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.369436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.369452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.369462 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.371399 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.383711 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.393342 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.402048 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.415367 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.423979 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.471598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.471642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.471651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.471669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.471679 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.573704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.573761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.573771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.573785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.573794 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.675864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.675904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.675914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.675927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.675937 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.778169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.778219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.778228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.778245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.778253 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.880278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.880318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.880326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.880341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.880348 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.982138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.982180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.982189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.982416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.982425 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.002547 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 15:17:40.095088858 +0000 UTC Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.049240 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.049262 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:41 crc kubenswrapper[4824]: E0121 11:11:41.049403 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:41 crc kubenswrapper[4824]: E0121 11:11:41.049503 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.084565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.084596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.084606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.084618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.084627 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.186612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.186642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.186651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.186665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.186675 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.242410 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/2.log" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.242841 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/1.log" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.244940 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" exitCode=1 Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.244983 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.245029 4824 scope.go:117] "RemoveContainer" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.245822 4824 scope.go:117] "RemoveContainer" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" Jan 21 11:11:41 crc kubenswrapper[4824]: E0121 11:11:41.246042 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.259163 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.269479 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.277303 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.286737 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.287930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.287975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.287984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.287997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.288005 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.294484 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.302591 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.310238 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.318335 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.326633 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.339374 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.350101 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.356700 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.363360 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.373379 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.381433 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.391848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.391883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.391895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.391913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.391923 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.394152 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.402485 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.494662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.494702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.494715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.494735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.494746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.596594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.596628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.596638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.596656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.596668 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.698495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.698526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.698534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.698548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.698558 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.800953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.801018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.801027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.801042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.801051 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.902935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.902999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.903010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.903026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.903037 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.003083 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 01:17:21.564019552 +0000 UTC Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.004532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.004560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.004571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.004583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.004592 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.048502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:42 crc kubenswrapper[4824]: E0121 11:11:42.048615 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.048686 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:42 crc kubenswrapper[4824]: E0121 11:11:42.048815 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.106430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.106454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.106462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.106473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.106481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.207822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.207849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.207858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.207868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.207875 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.248329 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/2.log" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.250494 4824 scope.go:117] "RemoveContainer" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" Jan 21 11:11:42 crc kubenswrapper[4824]: E0121 11:11:42.250618 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.259822 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.268021 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.279986 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.287570 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.294551 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.302765 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.309606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.309636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.309648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.309658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.309666 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.312416 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.321186 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.330257 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.342612 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.354824 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.364169 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.373481 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.382418 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.389784 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.399020 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.407941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.411681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.411710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.411720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.411734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.411744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.514485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.514542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.514551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.514564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.514574 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.616656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.616686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.616698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.616712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.616720 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.718326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.718357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.718365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.718377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.718387 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.819918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.819941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.819951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.819979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.819989 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.921348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.921369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.921378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.921390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.921416 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.004017 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 19:01:58.423277146 +0000 UTC Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.023999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.024064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.024076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.024101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.024405 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.048517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.048548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:43 crc kubenswrapper[4824]: E0121 11:11:43.048694 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:43 crc kubenswrapper[4824]: E0121 11:11:43.048739 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.126126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.126171 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.126181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.126197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.126207 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.229123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.229149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.229159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.229170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.229177 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.331482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.331523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.331537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.331553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.331565 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.434257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.434293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.434304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.434317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.434328 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.536872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.536921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.536931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.536948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.536973 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.639298 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.639326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.639336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.639347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.639382 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.741562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.741618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.741630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.741643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.741653 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.843608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.843663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.843675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.843694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.843707 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.945773 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.945821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.945830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.945847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.945857 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.004629 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 00:01:07.384168044 +0000 UTC Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048316 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048346 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:44 crc kubenswrapper[4824]: E0121 11:11:44.048455 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:44 crc kubenswrapper[4824]: E0121 11:11:44.048655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048765 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.056220 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:44 crc kubenswrapper[4824]: E0121 11:11:44.056382 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:44 crc kubenswrapper[4824]: E0121 11:11:44.056448 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:00.056431943 +0000 UTC m=+62.349461235 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.151035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.151071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.151081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.151095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.151105 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.252885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.252925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.252936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.252952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.252986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.354996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.355027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.355036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.355048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.355056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.457314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.457345 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.457354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.457365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.457376 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.559251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.559285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.559293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.559312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.559321 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.662395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.662426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.662435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.662448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.662457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.763940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.764034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.764046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.764071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.764086 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.866422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.866465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.866475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.866490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.866502 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.968686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.968726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.968736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.968749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.968759 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.004881 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 16:56:48.493862819 +0000 UTC Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.048647 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.048706 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:45 crc kubenswrapper[4824]: E0121 11:11:45.048774 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:45 crc kubenswrapper[4824]: E0121 11:11:45.049075 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.071507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.071544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.071556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.071571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.071581 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.174195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.174233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.174259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.174274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.174284 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.276757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.276792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.276800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.276814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.276822 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.379403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.379450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.379463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.379479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.379488 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.481108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.481145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.481153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.481166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.481175 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.583558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.583606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.583618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.583634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.583643 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.685929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.685994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.686007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.686033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.686043 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.788070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.788136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.788149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.788178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.788193 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.889723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.889762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.889771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.889786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.889795 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.991815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.991849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.991858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.991870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.991879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.005995 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 02:15:42.067088464 +0000 UTC Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.049270 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.049383 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.049276 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.049725 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.094146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.094174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.094184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.094195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.094205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.196173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.196223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.196233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.196245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.196264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.298524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.298563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.298572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.298585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.298593 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.400356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.400392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.400402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.400416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.400426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.501813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.501845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.501852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.501864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.501872 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.603553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.603578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.603586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.603597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.603606 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.705137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.705169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.705177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.705191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.705199 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.782168 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.782240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.782278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.782331 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.782333 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:12:18.782312245 +0000 UTC m=+81.075341547 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.782410 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:18.782389001 +0000 UTC m=+81.075418303 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.782421 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.782473 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:18.7824626 +0000 UTC m=+81.075491892 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.807454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.807500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.807510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.807520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.807529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.883183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.883240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883307 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883331 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883343 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883383 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:18.883371316 +0000 UTC m=+81.176400618 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883331 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883465 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883477 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883508 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:18.883499678 +0000 UTC m=+81.176528971 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.909266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.909291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.909299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.909310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.909318 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.006790 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 13:54:49.079161465 +0000 UTC Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.011054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.011085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.011094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.011106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.011115 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.048647 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.048692 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:47 crc kubenswrapper[4824]: E0121 11:11:47.048731 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:47 crc kubenswrapper[4824]: E0121 11:11:47.048798 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.112573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.112601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.112610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.112620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.112630 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.214285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.214315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.214323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.214336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.214344 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.316408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.316441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.316449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.316460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.316468 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.418463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.418497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.418505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.418521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.418529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.519699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.519733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.519741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.519752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.519761 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.621236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.621281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.621290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.621301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.621309 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.723108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.723141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.723150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.723161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.723169 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.825044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.825067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.825075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.825086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.825093 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.926509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.926542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.926551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.926562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.926570 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.007697 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 18:26:34.155619471 +0000 UTC Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.027691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.027717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.027725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.027735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.027744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.049231 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.049330 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.049388 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.049508 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.057557 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.064918 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.073668 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.081249 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.089815 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.101869 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.109531 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.116006 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.123720 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.129715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.129736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.129744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.129757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.129768 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.131189 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.139776 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.146824 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.155872 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.163125 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.171810 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.183649 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.190501 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.221080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.221114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.221140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.221166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.221174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.229111 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.231207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.231239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.231249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.231260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.231282 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.238793 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.240650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.240683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.240692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.240705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.240714 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.249735 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.251886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.251907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.251914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.251923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.251930 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.259503 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.261508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.261592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.261602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.261614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.261623 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.269903 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.270023 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.270838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.270862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.270872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.270880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.270887 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.372614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.372642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.372651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.372662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.372669 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.473971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.474003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.474012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.474024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.474031 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.576262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.576309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.576320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.576333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.576342 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.678087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.678123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.678131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.678143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.678152 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.779895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.779926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.779934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.779944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.779951 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.882140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.882178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.882187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.882198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.882205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.984026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.984073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.984085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.984099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.984109 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.008648 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 10:30:56.229177316 +0000 UTC Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.049194 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.049203 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:49 crc kubenswrapper[4824]: E0121 11:11:49.049302 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:49 crc kubenswrapper[4824]: E0121 11:11:49.049379 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.085680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.085715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.085727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.085738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.085746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.187680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.187709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.187718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.187729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.187738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.290207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.290251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.290260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.290272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.290295 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.392001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.392043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.392075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.392087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.392095 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.493784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.493823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.493831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.493843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.493854 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.596161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.596200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.596210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.596225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.596234 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.698556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.698591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.698600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.698613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.698623 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.800502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.800544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.800554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.800571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.800580 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.902575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.902630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.902641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.902665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.902679 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.005190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.005242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.005253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.005266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.005295 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.009415 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 01:38:54.221692997 +0000 UTC Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.048587 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:50 crc kubenswrapper[4824]: E0121 11:11:50.048703 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.048588 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:50 crc kubenswrapper[4824]: E0121 11:11:50.048892 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.107403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.107437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.107448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.107462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.107473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.208891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.208922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.208931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.208947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.209018 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.311562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.311598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.311608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.311625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.311635 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.413808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.414095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.414105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.414121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.414133 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.515451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.515485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.515493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.515505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.515514 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.617118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.617164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.617173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.617185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.617195 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.718414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.718452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.718463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.718476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.718485 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.820124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.820151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.820184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.820199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.820208 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.921704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.921740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.921749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.921759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.921767 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.009751 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 01:44:52.968773125 +0000 UTC Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.023329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.023370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.023380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.023394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.023403 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.048579 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.048606 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:51 crc kubenswrapper[4824]: E0121 11:11:51.048667 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:51 crc kubenswrapper[4824]: E0121 11:11:51.048737 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.125596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.125626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.125634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.125646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.125654 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.227391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.227422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.227432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.227445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.227456 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.329630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.329668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.329677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.329690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.329700 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.432242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.432287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.432311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.432327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.432337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.534950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.535028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.535038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.535055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.535063 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.637149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.637195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.637206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.637223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.637231 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.739261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.739351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.739372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.739399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.739412 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.841927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.841974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.841985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.841999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.842008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.944223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.944292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.944317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.944340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.944354 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.010930 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 03:06:13.665639909 +0000 UTC Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.046597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.046675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.046686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.046704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.046714 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.048927 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.048980 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:52 crc kubenswrapper[4824]: E0121 11:11:52.049046 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:52 crc kubenswrapper[4824]: E0121 11:11:52.049131 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.149020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.149090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.149103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.149123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.149134 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.253516 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.253554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.253563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.253575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.253583 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.355732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.355764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.355773 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.355785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.355792 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.457420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.457452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.457460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.457471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.457479 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.559765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.559794 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.559801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.559811 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.559818 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.662158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.662192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.662202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.662213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.662221 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.763614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.763636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.763645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.763655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.763663 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.865604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.865636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.865644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.865654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.865662 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.967484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.967513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.967521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.967531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.967541 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.012151 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 19:26:48.282007098 +0000 UTC Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.048836 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.048862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:53 crc kubenswrapper[4824]: E0121 11:11:53.048939 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:53 crc kubenswrapper[4824]: E0121 11:11:53.049023 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.049829 4824 scope.go:117] "RemoveContainer" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" Jan 21 11:11:53 crc kubenswrapper[4824]: E0121 11:11:53.050163 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.069484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.069524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.069538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.069553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.069566 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.171720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.171792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.171832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.171861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.171874 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.273868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.273899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.273909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.273923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.273950 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.375779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.375826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.375836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.375850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.375860 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.477202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.477235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.477243 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.477256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.477266 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.579574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.579634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.579647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.579675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.579688 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.682010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.682081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.682093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.682117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.682133 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.718426 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.727647 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.727670 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.737167 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.753081 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.760468 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.768748 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.776270 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.783310 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.784031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.784095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.784110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.784125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.784137 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.792480 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.801515 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.815660 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.824377 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.834757 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.842152 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.850429 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.860572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.870064 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.877879 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.886277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.886354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.886363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.886394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.886405 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.988156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.988189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.988198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.988211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.988219 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.012613 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 22:26:05.602798767 +0000 UTC Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.048395 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.048433 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:54 crc kubenswrapper[4824]: E0121 11:11:54.048527 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:54 crc kubenswrapper[4824]: E0121 11:11:54.048598 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.090445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.090478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.090486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.090498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.090506 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.192454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.192498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.192507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.192520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.192528 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.293952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.294000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.294008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.294020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.294028 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.395276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.395310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.395329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.395342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.395350 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.496504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.496527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.496534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.496544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.496551 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.598189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.598241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.598250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.598260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.598268 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.700655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.700722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.700737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.700761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.700775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.802527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.802559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.802569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.802581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.802589 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.904479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.904500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.904514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.904523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.904529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.006384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.006479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.006489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.006524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.006533 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.012736 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 19:19:19.242227906 +0000 UTC Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.049123 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:55 crc kubenswrapper[4824]: E0121 11:11:55.049201 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.049123 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:55 crc kubenswrapper[4824]: E0121 11:11:55.049283 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.108066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.108104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.108113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.108122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.108130 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.209068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.209096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.209104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.209114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.209126 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.310682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.310708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.310716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.310726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.310734 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.413086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.413134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.413144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.413155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.413163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.515041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.515070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.515079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.515089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.515096 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.616887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.616916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.616925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.616935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.616942 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.726220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.727058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.727066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.727078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.727086 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.828815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.828845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.828853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.828864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.828871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.930780 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.930814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.930823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.930847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.930855 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.013691 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 14:58:39.769297053 +0000 UTC Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.033090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.033117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.033142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.033154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.033162 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.048283 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.048305 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:56 crc kubenswrapper[4824]: E0121 11:11:56.048398 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:56 crc kubenswrapper[4824]: E0121 11:11:56.048449 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.135244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.135279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.135288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.135302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.135312 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.237601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.237636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.237647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.237660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.237669 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.339897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.339947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.339976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.339990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.339998 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.442297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.442348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.442361 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.442377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.442387 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.544272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.544303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.544311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.544321 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.544329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.646136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.646168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.646177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.646188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.646196 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.748358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.748391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.748399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.748409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.748420 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.850776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.850865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.850876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.850904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.850918 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.953765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.953807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.953816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.953832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.953842 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.014545 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 15:14:16.766274193 +0000 UTC Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.048813 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:57 crc kubenswrapper[4824]: E0121 11:11:57.048933 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.049024 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:57 crc kubenswrapper[4824]: E0121 11:11:57.049109 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.056567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.056602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.056612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.056626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.056635 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.158769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.158816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.158824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.158836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.158844 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.261158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.261195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.261204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.261217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.261228 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.363984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.364023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.364031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.364044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.364052 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.466188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.466217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.466227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.466239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.466246 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.568559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.568601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.568610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.568623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.568632 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.670691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.670726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.670735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.670747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.670756 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.773262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.773297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.773307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.773321 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.773329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.875315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.875369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.875377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.875388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.875397 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.976942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.976992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.977001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.977016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.977024 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.015075 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 04:16:54.091968613 +0000 UTC Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.049237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.049334 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.049552 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.049809 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.058792 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.066451 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.077780 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.078632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.078660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.078669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.078682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.078693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.090343 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.099560 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.107427 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.119879 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.129507 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.137389 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.145733 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.154895 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.162707 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.170863 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.177688 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.180284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.180336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.180346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.180371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.180380 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.185057 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.194062 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.206661 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.214620 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.282092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.282128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.282140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.282165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.282177 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.383982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.384116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.384182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.384250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.384307 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.486204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.486237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.486245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.486257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.486264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.588066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.588104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.588113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.588126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.588135 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.590639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.590669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.590677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.590690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.590698 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.600209 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.602438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.602460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.602467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.602477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.602484 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.611275 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.614000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.614092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.614158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.614220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.614284 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.622421 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.624706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.624784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.624793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.624804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.624811 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.633265 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.635292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.635417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.635478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.635542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.635599 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.643433 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.643927 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.690254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.690303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.690314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.690328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.690336 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.792110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.792166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.792176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.792210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.792223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.895051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.895106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.895116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.895131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.895139 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.996892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.996934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.996943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.996972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.996982 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.015599 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 04:06:58.95960934 +0000 UTC Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.049038 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.049078 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:59 crc kubenswrapper[4824]: E0121 11:11:59.049344 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:59 crc kubenswrapper[4824]: E0121 11:11:59.049233 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.099016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.099053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.099063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.099078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.099087 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.201103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.201286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.201374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.201453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.201520 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.303898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.304144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.304206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.304270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.304320 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.406665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.406742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.406754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.406779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.406793 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.509467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.509526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.509541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.509565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.509577 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.611567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.611611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.611620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.611634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.611642 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.713702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.713741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.713749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.713763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.713774 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.816061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.816108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.816118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.816133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.816142 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.917782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.917816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.917827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.917843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.917852 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.016432 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 01:48:51.567181656 +0000 UTC Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.020003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.020046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.020058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.020076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.020096 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.049115 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:00 crc kubenswrapper[4824]: E0121 11:12:00.049220 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.049253 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:00 crc kubenswrapper[4824]: E0121 11:12:00.049316 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.104415 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:00 crc kubenswrapper[4824]: E0121 11:12:00.104508 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:12:00 crc kubenswrapper[4824]: E0121 11:12:00.104558 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:32.104543511 +0000 UTC m=+94.397572813 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.121747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.121775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.121786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.121798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.121807 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.224309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.224353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.224365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.224395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.224405 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.326599 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.326643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.326653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.326669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.326683 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.429192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.429242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.429252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.429266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.429301 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.531751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.531786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.531793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.531804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.531811 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.633017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.633051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.633062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.633074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.633082 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.735028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.735050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.735059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.735070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.735077 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.836798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.836840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.836848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.836862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.836870 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.939220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.939264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.939274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.939289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.939301 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.017236 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 02:45:04.660901158 +0000 UTC Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.041317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.041475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.041536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.041601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.041671 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.048680 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.048841 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:01 crc kubenswrapper[4824]: E0121 11:12:01.048862 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:01 crc kubenswrapper[4824]: E0121 11:12:01.049075 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.143641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.143686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.143695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.143711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.143720 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.245450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.245497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.245507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.245522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.245531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.347879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.347939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.347950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.347991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.348005 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.450633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.450705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.450718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.450768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.450783 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.555809 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.555936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.556051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.556129 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.556188 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.658093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.658206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.658271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.658331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.658411 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.760235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.760296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.760306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.760325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.760336 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.862711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.862774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.862787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.862814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.862827 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.964913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.964977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.964989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.965006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.965018 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.018277 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 22:51:06.852216222 +0000 UTC Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.048831 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.048862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:02 crc kubenswrapper[4824]: E0121 11:12:02.048999 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:02 crc kubenswrapper[4824]: E0121 11:12:02.049137 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.067080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.067116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.067127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.067141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.067152 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.169796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.169860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.169874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.169901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.169917 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.272084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.272117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.272126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.272137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.272146 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.373967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.374001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.374011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.374022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.374031 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.475194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.475223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.475232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.475242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.475250 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.577336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.577443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.577458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.577482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.577494 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.679000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.679033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.679043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.679054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.679062 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.780554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.780651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.780724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.780791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.780846 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.882403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.882436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.882447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.882459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.882471 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.984590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.984618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.984626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.984637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.984645 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.018856 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 17:33:40.855594493 +0000 UTC Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.048848 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.048863 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:03 crc kubenswrapper[4824]: E0121 11:12:03.048946 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:03 crc kubenswrapper[4824]: E0121 11:12:03.049037 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.086109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.086134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.086143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.086154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.086163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.188077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.188103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.188114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.188124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.188133 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.289682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.289713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.289723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.289739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.289751 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.296601 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/0.log" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.296637 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" containerID="75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15" exitCode=1 Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.296666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerDied","Data":"75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.296944 4824 scope.go:117] "RemoveContainer" containerID="75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.307030 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.321752 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.336058 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.346591 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.357517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.365444 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.378922 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.388567 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.391926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.391979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.391991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.392010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.392024 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.398570 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.408676 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.418424 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.426673 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.436921 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.444709 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.452886 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.461530 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.478360 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.488158 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.494839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.494879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.494889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.494906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.494918 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.597179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.597213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.597432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.597450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.597459 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.699334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.699376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.699389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.699421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.699434 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.801819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.801868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.801879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.801894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.801903 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.903472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.903505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.903518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.903532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.903540 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.005802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.005830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.005839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.005852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.005860 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.019278 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 22:44:58.881872006 +0000 UTC Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.048634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.048634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:04 crc kubenswrapper[4824]: E0121 11:12:04.048789 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:04 crc kubenswrapper[4824]: E0121 11:12:04.048837 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.107618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.107653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.107661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.107686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.107697 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.210075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.210128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.210137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.210150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.210159 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.302234 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/0.log" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.302296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerStarted","Data":"49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.311804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.311890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.311908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.311933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.311947 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.316503 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.326825 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.335171 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.345471 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.353425 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.363988 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.375853 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.386973 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.395551 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.404611 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.413698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.413748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.413761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.413775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.413782 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.414089 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.430725 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.440791 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.452012 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.460485 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.469890 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.481149 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.491911 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.517020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.517085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.517100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.517119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.517129 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.620006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.620073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.620087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.620111 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.620124 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.721799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.721935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.722036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.722116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.722174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.824680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.824729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.824740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.824758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.824767 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.926504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.926556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.926565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.926579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.926590 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.019837 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 06:58:50.036087652 +0000 UTC Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.028077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.028111 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.028121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.028135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.028143 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.048225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:05 crc kubenswrapper[4824]: E0121 11:12:05.048419 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.048270 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:05 crc kubenswrapper[4824]: E0121 11:12:05.048634 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.130552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.130589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.130618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.130632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.130639 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.232644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.232690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.232702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.232722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.232738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.335558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.335594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.335612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.335626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.335637 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.438525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.438564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.438574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.438588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.438598 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.540988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.541055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.541068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.541096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.541109 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.644007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.644058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.644069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.644091 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.644106 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.746503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.746556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.746567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.746584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.746593 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.848593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.848642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.848654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.848669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.848681 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.949832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.950055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.950127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.950235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.950315 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.020750 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 17:24:25.064947226 +0000 UTC Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.049125 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:06 crc kubenswrapper[4824]: E0121 11:12:06.049236 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.049324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:06 crc kubenswrapper[4824]: E0121 11:12:06.049599 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.052276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.052317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.052327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.052340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.052349 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.154738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.154788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.154798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.154815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.154825 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.256768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.256801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.256813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.256826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.256837 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.359063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.359100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.359109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.359124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.359134 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.461160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.461199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.461209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.461223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.461233 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.563267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.563307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.563318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.563330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.563342 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.665626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.665663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.665672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.665687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.665696 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.767307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.767347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.767356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.767374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.767385 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.869079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.869119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.869134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.869151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.869163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.970942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.970999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.971012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.971029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.971040 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.021460 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 02:46:11.749001905 +0000 UTC Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.048852 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.048920 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:07 crc kubenswrapper[4824]: E0121 11:12:07.049148 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:07 crc kubenswrapper[4824]: E0121 11:12:07.049258 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.049344 4824 scope.go:117] "RemoveContainer" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.073458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.073495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.073505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.073520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.073531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.175872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.175904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.175914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.175928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.175938 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.278142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.278190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.278199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.278219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.278230 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.311433 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/2.log" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.313518 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.314524 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.333741 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.343243 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.359894 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.379714 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.381307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.381343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.381355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.381373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.381383 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.397195 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.404901 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.418592 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.427433 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.440932 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.449564 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.457894 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.466971 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.476594 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.484122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.484170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.484181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.484199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.484210 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.486543 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.495511 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.502777 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.510132 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.521010 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.586519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.586565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.586578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.586597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.586609 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.688841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.688900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.688911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.688929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.688944 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.790377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.790413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.790435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.790449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.790460 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.892679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.892713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.892722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.892736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.892746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.994392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.994435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.994446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.994463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.994473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.022054 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 08:15:39.520093142 +0000 UTC Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.048274 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.048281 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.048383 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.048465 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.060564 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.069941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.076637 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.086612 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095568 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095918 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.103697 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.116646 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.127381 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.137147 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.144289 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.152625 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.160620 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.169017 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.176713 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.185071 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.192857 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.197059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.197081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.197091 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.197103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.197112 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.205469 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.213009 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.299045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.299076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.299087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.299099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.299107 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.317309 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/3.log" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.317987 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/2.log" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.320054 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" exitCode=1 Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.320084 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.320112 4824 scope.go:117] "RemoveContainer" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.320613 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.320779 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.332183 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.340632 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.349341 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.357211 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.366624 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.373335 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.380563 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.388262 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.400464 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:07Z\\\",\\\"message\\\":\\\"UUIDName:}]\\\\nI0121 11:12:07.749346 6924 services_controller.go:453] Built service openshift-console/console template LB for network=default: []services.LB{}\\\\nI0121 11:12:07.749353 6924 services_controller.go:454] Service openshift-console/console for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0121 11:12:07.747279 6924 services_controller.go:434] Service openshift-kube-storage-version-migrator-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-storage-version-migrator-operator e1639a86-fb7f-46de-9d5e-4aee16dccea1 4372 0 2025-02-23 05:12:25 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-storage-version-migrator-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007778417 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:htt\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.401598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.401624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.401633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.401645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.401653 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.407544 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.414315 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.423354 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.430909 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.438762 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.447244 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.453559 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.466050 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.474267 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.504170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.504199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.504209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.504225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.504234 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.606530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.606564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.606573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.606586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.606595 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.697791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.697840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.697852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.697866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.697876 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.708724 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.711741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.711789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.711803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.711821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.711830 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.721278 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.724475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.724507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.724518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.724532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.724542 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.733855 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.737561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.737600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.737610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.737626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.737638 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.747256 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.749612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.749636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.749648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.749659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.749667 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.760142 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.760253 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.761596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.761634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.761644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.761675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.761687 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.863842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.863891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.863901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.863917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.863930 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.966363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.966396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.966408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.966437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.966450 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.022317 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 20:24:07.499506491 +0000 UTC Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.048555 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.048692 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:09 crc kubenswrapper[4824]: E0121 11:12:09.048799 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:09 crc kubenswrapper[4824]: E0121 11:12:09.048928 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.067893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.067922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.067933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.067945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.067977 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.170730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.170766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.170777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.170789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.170798 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.272921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.272950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.272976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.272992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.273003 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.324633 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/3.log" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.328307 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:09 crc kubenswrapper[4824]: E0121 11:12:09.328462 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.344835 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.354983 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.364347 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.373596 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.375912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.376001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.376017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.376038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.376053 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.385805 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.396500 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.407529 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.415605 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.423533 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.432798 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.446048 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:07Z\\\",\\\"message\\\":\\\"UUIDName:}]\\\\nI0121 11:12:07.749346 6924 services_controller.go:453] Built service openshift-console/console template LB for network=default: []services.LB{}\\\\nI0121 11:12:07.749353 6924 services_controller.go:454] Service openshift-console/console for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0121 11:12:07.747279 6924 services_controller.go:434] Service openshift-kube-storage-version-migrator-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-storage-version-migrator-operator e1639a86-fb7f-46de-9d5e-4aee16dccea1 4372 0 2025-02-23 05:12:25 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-storage-version-migrator-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007778417 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:htt\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.453481 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.462990 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.472117 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.478353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.478388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.478399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.478429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.478441 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.480016 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.488573 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.496329 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.503053 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.581796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.581844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.581864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.581888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.581911 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.684329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.684358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.684369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.684384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.684394 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.786210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.786262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.786276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.786296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.786310 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.888702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.888748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.888759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.888777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.888786 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.990712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.990744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.990752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.990766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.990775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.022594 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 02:45:45.467635292 +0000 UTC Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.048324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:10 crc kubenswrapper[4824]: E0121 11:12:10.048438 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.048468 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:10 crc kubenswrapper[4824]: E0121 11:12:10.048565 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.093040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.093149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.093221 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.093275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.093326 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.195312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.195522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.195586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.195642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.195692 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.297481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.297551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.297566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.297591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.297603 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.400241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.400440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.400624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.400780 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.400912 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.503276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.503322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.503334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.503351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.503363 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.606238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.606370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.606680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.606756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.606824 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.708152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.708178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.708187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.708200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.708209 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.809549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.809581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.809590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.809605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.809614 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.910701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.910734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.910743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.910756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.910764 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.012371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.012427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.012436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.012449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.012457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.023708 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 22:47:03.820783195 +0000 UTC Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.049129 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.049286 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:11 crc kubenswrapper[4824]: E0121 11:12:11.049333 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:11 crc kubenswrapper[4824]: E0121 11:12:11.049554 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.114823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.114867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.114877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.114888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.114898 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.217498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.217543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.217553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.217567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.217578 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.319363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.319423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.319433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.319449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.319461 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.421493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.421544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.421553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.421566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.421576 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.523674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.523717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.523726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.523741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.523751 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.625189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.625219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.625246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.625260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.625266 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.727760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.727788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.727797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.727808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.727815 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.829585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.829625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.829636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.829653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.829664 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.931453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.931494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.931503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.931520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.931528 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.024762 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 19:09:19.15886122 +0000 UTC Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.033365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.033389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.033399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.033419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.033427 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.048973 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.048991 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:12 crc kubenswrapper[4824]: E0121 11:12:12.049082 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:12 crc kubenswrapper[4824]: E0121 11:12:12.049215 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.135398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.135464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.135487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.135500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.135511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.237012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.237041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.237050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.237064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.237075 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.338679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.338729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.338741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.338759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.338768 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.440774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.440805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.440814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.440825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.440834 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.542804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.542839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.542847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.542860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.542868 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.644675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.644730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.644740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.644757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.644766 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.746332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.746360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.746369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.746380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.746390 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.848101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.848139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.848150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.848165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.848174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.949865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.949899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.949907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.949920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.949928 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.025560 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 22:38:28.950155576 +0000 UTC Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.049191 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.049211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:13 crc kubenswrapper[4824]: E0121 11:12:13.049281 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:13 crc kubenswrapper[4824]: E0121 11:12:13.049358 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.051278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.051300 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.051311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.051322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.051331 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.152486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.152512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.152520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.152531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.152538 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.256163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.256194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.256202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.256214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.256222 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.358037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.358071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.358081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.358094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.358103 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.460093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.460121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.460128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.460137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.460144 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.562250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.562281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.562289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.562299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.562306 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.663846 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.663876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.663883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.663909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.663918 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.765580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.765608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.765616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.765628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.765635 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.867831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.867860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.867868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.867878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.867886 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.971278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.971306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.971314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.971327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.971337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.025977 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 00:29:47.057543229 +0000 UTC Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.048505 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.048584 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:14 crc kubenswrapper[4824]: E0121 11:12:14.048665 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:14 crc kubenswrapper[4824]: E0121 11:12:14.048728 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.056510 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.073472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.073499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.073507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.073517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.073525 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.175619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.175650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.175658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.175669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.175677 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.277860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.277895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.277903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.277916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.277924 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.379783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.379827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.379835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.379846 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.379853 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.481669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.481698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.481706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.481718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.481730 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.583596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.583636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.583645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.583673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.583680 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.685546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.685583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.685592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.685605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.685613 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.787228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.787273 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.787283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.787296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.787304 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.889510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.889544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.889569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.889583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.889592 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.991124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.991162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.991170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.991184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.991193 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.026683 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 00:19:52.153953239 +0000 UTC Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.049221 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.049245 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:15 crc kubenswrapper[4824]: E0121 11:12:15.049326 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:15 crc kubenswrapper[4824]: E0121 11:12:15.049391 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.093027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.093058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.093068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.093082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.093090 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.195114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.195150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.195158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.195174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.195182 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.296646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.296699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.296708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.296722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.296732 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.398165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.398199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.398207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.398218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.398227 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.499414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.499446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.499455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.499468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.499476 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.600727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.600755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.600765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.600777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.600785 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.702237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.702268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.702277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.702306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.702316 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.806731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.806767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.806777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.806789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.806797 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.908015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.908050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.908059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.908073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.908081 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.010068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.010108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.010119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.010133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.010141 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.027420 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 09:48:22.72384722 +0000 UTC Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.048778 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:16 crc kubenswrapper[4824]: E0121 11:12:16.048883 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.048914 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:16 crc kubenswrapper[4824]: E0121 11:12:16.049041 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.112028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.112054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.112063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.112082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.112091 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.214153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.214191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.214202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.214217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.214225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.316442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.316470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.316478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.316491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.316501 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.418382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.418431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.418439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.418452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.418461 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.520043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.520074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.520082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.520094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.520103 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.622003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.622048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.622056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.622068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.622077 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.723710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.723740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.723750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.723763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.723772 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.825449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.825480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.825489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.825501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.825510 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.927004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.927034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.927042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.927054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.927061 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.027997 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 08:58:56.390507961 +0000 UTC Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.029301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.029339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.029347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.029359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.029369 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.048579 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.048601 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:17 crc kubenswrapper[4824]: E0121 11:12:17.048684 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:17 crc kubenswrapper[4824]: E0121 11:12:17.048760 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.131334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.131367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.131375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.131402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.131412 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.233226 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.233259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.233268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.233283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.233291 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.335173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.335207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.335217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.335231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.335239 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.436997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.437035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.437046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.437058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.437067 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.539156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.539187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.539196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.539209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.539218 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.641472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.641524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.641535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.641548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.641558 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.743800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.743833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.743841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.743854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.743862 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.845556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.845608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.845619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.845631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.845640 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.947479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.947518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.947527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.947540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.947550 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.028278 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 19:47:14.845753186 +0000 UTC Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.048511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.048548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.048888 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.049000 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.049774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.050257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.050287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.050301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.050309 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.058769 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.066229 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44f8a34-1578-4634-b62e-8e222844b397\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd58f2438b6f9cbb2c37687e39f02cdf96a029477b1ae33f6dc9f3b70ecc926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.075735 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.083866 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.092524 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.101943 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.108801 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.121714 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.129512 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.137797 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.145396 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.151934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.151982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.151992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.152004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.152013 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.153479 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.161067 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.170782 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.177928 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.185300 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.193555 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.205805 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:07Z\\\",\\\"message\\\":\\\"UUIDName:}]\\\\nI0121 11:12:07.749346 6924 services_controller.go:453] Built service openshift-console/console template LB for network=default: []services.LB{}\\\\nI0121 11:12:07.749353 6924 services_controller.go:454] Service openshift-console/console for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0121 11:12:07.747279 6924 services_controller.go:434] Service openshift-kube-storage-version-migrator-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-storage-version-migrator-operator e1639a86-fb7f-46de-9d5e-4aee16dccea1 4372 0 2025-02-23 05:12:25 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-storage-version-migrator-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007778417 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:htt\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.213535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.253838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.253870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.253878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.253893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.253901 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.355206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.355241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.355250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.355261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.355269 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.457303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.457352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.457360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.457376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.457398 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.559606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.559638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.559647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.559677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.559695 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.662138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.662190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.662199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.662214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.662223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.764120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.764152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.764160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.764172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.764181 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.866362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.866422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.866433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.866445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.866452 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.875795 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.875920 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:22.875904663 +0000 UTC m=+145.168933955 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.875994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.876016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.876077 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.876119 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:13:22.876111963 +0000 UTC m=+145.169141255 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.876276 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.876422 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:13:22.876406517 +0000 UTC m=+145.169435810 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.968140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.968398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.968480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.968550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.968630 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.976512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.976556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.976653 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.976674 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.976685 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.976715 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:13:22.976704539 +0000 UTC m=+145.269733831 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.976912 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.977007 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.977070 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.977162 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:13:22.977149426 +0000 UTC m=+145.270178718 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.001601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.001637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.001645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.001662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.001671 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.010511 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.013160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.013192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.013200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.013212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.013220 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.023117 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.025917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.025970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.025980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.025993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.026001 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.028788 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 15:37:34.072729343 +0000 UTC Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.034032 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.036376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.036418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.036427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.036441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.036449 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.044767 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.047074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.047099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.047108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.047137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.047146 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.048225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.048245 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.048326 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.048396 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.056004 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.056117 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.070340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.070372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.070391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.070405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.070414 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.172753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.172996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.173117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.173205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.173285 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.275586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.275855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.275948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.276043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.276126 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.377626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.377844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.377943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.378057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.378203 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.479897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.480123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.480132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.480144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.480153 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.581365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.581393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.581401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.581410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.581419 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.682692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.682803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.682879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.682971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.683135 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.785336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.785368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.785388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.785400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.785409 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.887481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.887512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.887519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.887533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.887541 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.989710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.989742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.989750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.989764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.989772 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.029135 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 02:02:09.851170023 +0000 UTC Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.048403 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.048454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:20 crc kubenswrapper[4824]: E0121 11:12:20.048501 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:20 crc kubenswrapper[4824]: E0121 11:12:20.048545 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.049186 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:20 crc kubenswrapper[4824]: E0121 11:12:20.049316 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.090916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.090943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.090951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.090981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.090991 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.192883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.192913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.192921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.192932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.192941 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.294697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.294724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.294732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.294742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.294749 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.396170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.396198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.396206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.396216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.396223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.498044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.498098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.498107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.498118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.498126 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.599629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.599656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.599663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.599674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.599683 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.701692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.701723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.701732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.701744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.701754 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.802884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.802925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.802937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.802982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.802993 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.905076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.905105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.905134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.905146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.905154 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.006764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.006796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.006806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.006818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.006846 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.030174 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 15:21:32.155937181 +0000 UTC Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.048451 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:21 crc kubenswrapper[4824]: E0121 11:12:21.048528 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.048616 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:21 crc kubenswrapper[4824]: E0121 11:12:21.048746 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.108697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.108734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.108744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.108757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.108766 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.209927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.209969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.209978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.209988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.209996 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.312188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.312841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.312988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.313122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.313187 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.415701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.415913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.415996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.416062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.416132 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.518421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.518664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.518754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.518820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.518880 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.621575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.621609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.621617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.621632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.621641 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.723435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.723471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.723481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.723501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.723509 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.825160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.825193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.825203 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.825217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.825225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.927611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.927869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.927953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.928050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.928113 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.029900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.029933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.029942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.029977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.029986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.030256 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 13:55:53.470736332 +0000 UTC Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.048479 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.048484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:22 crc kubenswrapper[4824]: E0121 11:12:22.048773 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:22 crc kubenswrapper[4824]: E0121 11:12:22.048853 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.132049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.132081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.132093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.132104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.132112 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.233635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.233666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.233673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.233684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.233693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.335787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.335813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.335820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.335832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.335839 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.437607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.437658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.437667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.437682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.437690 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.539739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.539781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.539790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.539804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.539813 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.641409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.641447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.641457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.641471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.641480 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.743084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.743122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.743131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.743145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.743154 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.844832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.844875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.844884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.844895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.844903 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.946519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.946547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.946559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.946570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.946579 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.031150 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 20:58:29.913446283 +0000 UTC Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048269 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048316 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:23 crc kubenswrapper[4824]: E0121 11:12:23.048423 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:23 crc kubenswrapper[4824]: E0121 11:12:23.048495 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048796 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.150351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.150398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.150406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.150418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.150426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.252389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.252421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.252429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.252441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.252449 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.354830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.354901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.354910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.354921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.354929 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.457033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.457062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.457070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.457080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.457090 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.558534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.558564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.558571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.558582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.558590 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.660554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.660592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.660600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.660668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.660676 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.762837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.762887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.762897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.762909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.762917 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.865381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.865431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.865440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.865453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.865461 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.967767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.967798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.967807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.967819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.967827 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.031689 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 08:59:56.766731956 +0000 UTC Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.048986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.049058 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:24 crc kubenswrapper[4824]: E0121 11:12:24.049109 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:24 crc kubenswrapper[4824]: E0121 11:12:24.049240 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.070194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.070232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.070242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.070256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.070266 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.171893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.171919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.171927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.171937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.171945 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.273886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.273926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.273935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.273949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.273976 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.376211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.376264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.376275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.376291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.376306 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.477741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.477793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.477803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.477819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.477829 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.579626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.579654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.579662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.579673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.579682 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.681239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.681283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.681292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.681307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.681316 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.782625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.782660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.782667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.782679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.782687 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.884671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.884708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.884717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.884732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.884740 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.986819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.986853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.986862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.986876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.986885 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.032561 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 05:37:05.730617272 +0000 UTC Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.048788 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.048788 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:25 crc kubenswrapper[4824]: E0121 11:12:25.048882 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:25 crc kubenswrapper[4824]: E0121 11:12:25.048931 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.088337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.088373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.088382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.088391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.088398 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.189889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.189919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.189928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.189936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.189943 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.291494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.291537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.291547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.291560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.291570 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.393432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.393467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.393477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.393488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.393498 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.494636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.494680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.494688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.494702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.494711 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.596477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.596502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.596509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.596519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.596527 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.698439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.698469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.698478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.698489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.698496 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.800140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.800171 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.800179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.800188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.800195 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.902068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.902096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.902104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.902114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.902121 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.003451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.003481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.003489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.003500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.003525 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.032940 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 19:58:06.973520129 +0000 UTC Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.049254 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:26 crc kubenswrapper[4824]: E0121 11:12:26.049378 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.049464 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:26 crc kubenswrapper[4824]: E0121 11:12:26.049617 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.105286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.105320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.105329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.105342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.105353 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.207028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.207065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.207073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.207084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.207092 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.309552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.309597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.309607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.309621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.309631 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.411553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.411586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.411594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.411608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.411618 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.513002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.513038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.513048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.513062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.513070 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.614674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.614731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.614740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.614755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.614764 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.716467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.716495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.716503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.716513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.716521 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.818033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.818066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.818075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.818087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.818095 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.920216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.920253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.920262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.920275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.920283 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.022137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.022172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.022180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.022194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.022202 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.033397 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 13:34:16.597437611 +0000 UTC Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.048874 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.048893 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:27 crc kubenswrapper[4824]: E0121 11:12:27.049156 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:27 crc kubenswrapper[4824]: E0121 11:12:27.049069 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.124015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.124049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.124057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.124068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.124076 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.226369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.226404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.226412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.226424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.226432 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.328192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.328228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.328238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.328251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.328258 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.429700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.429735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.429745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.429758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.429767 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.531797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.531830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.531840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.531851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.531859 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.633557 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.633589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.633600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.633613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.633622 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.735686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.735726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.735735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.735751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.735759 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.837379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.837411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.837420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.837432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.837439 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.938832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.938854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.938862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.938871 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.938879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.033758 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 06:59:54.159032978 +0000 UTC Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.040944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.040983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.040992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.041001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.041009 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.048321 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:28 crc kubenswrapper[4824]: E0121 11:12:28.048423 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.048441 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:28 crc kubenswrapper[4824]: E0121 11:12:28.048499 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.062533 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.071015 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.077971 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.085684 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.093032 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.103398 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.110653 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.119371 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.126084 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.133804 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.142459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.142498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.142509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.142523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.142533 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.145458 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:07Z\\\",\\\"message\\\":\\\"UUIDName:}]\\\\nI0121 11:12:07.749346 6924 services_controller.go:453] Built service openshift-console/console template LB for network=default: []services.LB{}\\\\nI0121 11:12:07.749353 6924 services_controller.go:454] Service openshift-console/console for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0121 11:12:07.747279 6924 services_controller.go:434] Service openshift-kube-storage-version-migrator-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-storage-version-migrator-operator e1639a86-fb7f-46de-9d5e-4aee16dccea1 4372 0 2025-02-23 05:12:25 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-storage-version-migrator-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007778417 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:htt\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.152280 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.158060 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.163747 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.169577 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44f8a34-1578-4634-b62e-8e222844b397\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd58f2438b6f9cbb2c37687e39f02cdf96a029477b1ae33f6dc9f3b70ecc926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.178553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.186212 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.193224 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.200574 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.244812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.244841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.244850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.244862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.244870 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.346013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.346041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.346051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.346062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.346072 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.447612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.447643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.447652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.447664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.447674 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.549748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.549777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.549785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.549796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.549805 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.651450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.651479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.651487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.651498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.651507 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.753212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.753236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.753244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.753317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.753329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.854949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.854997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.855005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.855015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.855023 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.956544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.956573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.956584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.956593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.956601 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.034786 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 20:56:25.353670252 +0000 UTC Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.049185 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.049214 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.049274 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.049330 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.058075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.058095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.058103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.058112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.058119 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.159858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.159883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.159891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.159900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.159908 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.226107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.226136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.226146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.226157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.226166 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.235163 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:29Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.237843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.237865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.237873 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.237884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.237891 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.245375 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:29Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.247426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.247447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.247456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.247464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.247472 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.254715 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:29Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.257011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.257035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.257043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.257051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.257058 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.264390 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:29Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.266303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.266325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.266333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.266341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.266359 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.273473 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:29Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.273592 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.274459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.274484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.274493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.274503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.274511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.376411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.376433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.376440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.376450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.376457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.478225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.478248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.478256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.478266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.478274 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.580378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.580405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.580413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.580424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.580431 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.682070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.682098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.682108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.682116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.682124 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.783745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.783765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.783774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.783782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.783790 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.884809 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.884834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.884841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.884851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.884858 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.986902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.986927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.986937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.986948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.986972 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.035409 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 20:29:34.881837332 +0000 UTC Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.049089 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.049139 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:30 crc kubenswrapper[4824]: E0121 11:12:30.049187 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:30 crc kubenswrapper[4824]: E0121 11:12:30.049248 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.088159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.088179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.088187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.088196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.088203 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.190534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.190565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.190572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.190580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.190587 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.292606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.292625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.292634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.292643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.292650 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.393841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.393906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.393916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.393929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.393937 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.496085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.496120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.496130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.496142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.496151 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.597578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.597604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.597612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.597622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.597629 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.699176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.699202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.699211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.699220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.699228 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.800501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.800534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.800542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.800554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.800563 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.903916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.904082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.904147 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.904175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.904184 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.006231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.006267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.006278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.006291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.006299 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.035513 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 03:46:03.149802342 +0000 UTC Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.048839 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:31 crc kubenswrapper[4824]: E0121 11:12:31.048925 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.048844 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:31 crc kubenswrapper[4824]: E0121 11:12:31.049036 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.108311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.108357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.108365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.108377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.108386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.210253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.210284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.210292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.210305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.210314 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.312442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.312474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.312482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.312493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.312501 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.414533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.414564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.414583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.414595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.414604 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.516311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.516410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.516419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.516428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.516434 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.617788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.617813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.617821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.617829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.617836 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.719505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.719532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.719540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.719549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.719556 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.821650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.821669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.821677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.821686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.821693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.923202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.923222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.923230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.923238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.923245 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.024919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.024935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.024943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.024974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.024983 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.036392 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 08:22:51.64614733 +0000 UTC Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.048805 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.049006 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:32 crc kubenswrapper[4824]: E0121 11:12:32.049131 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.049230 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:32 crc kubenswrapper[4824]: E0121 11:12:32.049252 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:32 crc kubenswrapper[4824]: E0121 11:12:32.049385 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.126818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.126842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.126849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.126860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.126866 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.179586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:32 crc kubenswrapper[4824]: E0121 11:12:32.179665 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:12:32 crc kubenswrapper[4824]: E0121 11:12:32.179727 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:13:36.179713996 +0000 UTC m=+158.472743288 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.229469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.229507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.229515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.229697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.229727 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.332992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.333035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.333044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.333060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.333070 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.435855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.435893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.435922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.435943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.435971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.538417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.538479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.538489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.538509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.538519 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.641548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.641594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.641604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.641620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.641627 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.744218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.744255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.744264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.744279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.744287 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.846415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.846483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.846498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.846520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.846531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.948997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.949038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.949047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.949061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.949070 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.037174 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 12:45:29.05598313 +0000 UTC Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.048518 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.048549 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:33 crc kubenswrapper[4824]: E0121 11:12:33.048677 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:33 crc kubenswrapper[4824]: E0121 11:12:33.048760 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.050692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.050733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.050741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.050759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.050768 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.152648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.152682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.152690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.152707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.152715 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.254796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.254832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.254841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.254854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.254863 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.356821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.356860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.356869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.356884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.356894 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.458384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.458420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.458429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.458445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.458454 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.560308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.560362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.560373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.560394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.560405 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.662420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.662479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.662487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.662522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.662531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.763870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.763904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.763912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.763926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.763935 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.865926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.866116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.866127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.866143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.866153 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.968650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.968703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.968713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.968730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.968738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.037584 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 05:58:41.730303231 +0000 UTC Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.048953 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.049053 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:34 crc kubenswrapper[4824]: E0121 11:12:34.049098 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:34 crc kubenswrapper[4824]: E0121 11:12:34.049193 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.070991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.071017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.071026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.071036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.071044 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.173305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.173347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.173356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.173368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.173376 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.275681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.275715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.275723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.275736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.275744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.377644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.377680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.377689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.377705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.377716 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.479320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.479374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.479383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.479399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.479408 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.581620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.581674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.581683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.581694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.581702 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.683540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.683585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.683594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.683609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.683618 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.785646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.785685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.785694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.785709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.785719 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.888047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.888082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.888090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.888105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.888113 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.989671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.989708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.989719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.989733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.989742 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.038400 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 09:30:56.989879119 +0000 UTC Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.048670 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.048729 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:35 crc kubenswrapper[4824]: E0121 11:12:35.048790 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:35 crc kubenswrapper[4824]: E0121 11:12:35.048918 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.091693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.091725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.091732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.091745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.091755 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.193902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.193933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.193941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.193977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.193986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.295858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.295896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.295906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.295919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.295927 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.397262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.397295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.397304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.397317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.397339 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.499648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.499686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.499709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.499723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.499731 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.601901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.601939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.601948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.601978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.601987 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.704642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.704678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.704687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.704701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.704709 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.806992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.807027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.807036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.807048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.807056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.908755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.908789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.908799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.908813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.908822 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.013940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.013996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.014005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.014021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.014030 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.050280 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 11:01:25.456081842 +0000 UTC Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.050335 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:36 crc kubenswrapper[4824]: E0121 11:12:36.050443 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.050468 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:36 crc kubenswrapper[4824]: E0121 11:12:36.050549 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.115855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.115892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.115899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.115913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.115922 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.218096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.218133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.218141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.218154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.218163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.319801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.319837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.319847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.319861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.319870 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.421413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.421452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.421460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.421474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.421481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.523705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.523745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.523754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.523770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.523779 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.625670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.625720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.625730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.625742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.625750 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.726844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.726877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.726886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.726900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.726909 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.828653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.828691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.828699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.828713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.828723 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.930415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.930451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.930461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.930473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.930482 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.032521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.032560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.032569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.032584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.032592 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.048864 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.048923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:37 crc kubenswrapper[4824]: E0121 11:12:37.049019 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:37 crc kubenswrapper[4824]: E0121 11:12:37.049089 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.051129 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 19:12:19.471814047 +0000 UTC Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.134646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.134680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.134688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.134704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.134713 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.236715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.236759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.236788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.236805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.236815 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.338429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.338466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.338476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.338490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.338498 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.440005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.440040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.440048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.440062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.440070 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.542150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.542188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.542197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.542211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.542219 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.644381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.644414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.644423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.644437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.644446 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.746706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.747290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.747304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.748032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.748051 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.849788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.849822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.849831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.849847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.849858 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.951715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.951743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.951751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.951762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.951772 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.048363 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:38 crc kubenswrapper[4824]: E0121 11:12:38.048493 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.048549 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:38 crc kubenswrapper[4824]: E0121 11:12:38.048842 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.051261 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 15:21:44.248961456 +0000 UTC Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.053707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.053747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.053756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.053766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.053775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.057104 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.064170 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.070715 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44f8a34-1578-4634-b62e-8e222844b397\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd58f2438b6f9cbb2c37687e39f02cdf96a029477b1ae33f6dc9f3b70ecc926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.079309 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.088735 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.096256 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.103653 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.115812 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.123462 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.129771 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.155643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.155676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.155685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.155700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.155709 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.163201 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-x94nb" podStartSLOduration=84.163189144 podStartE2EDuration="1m24.163189144s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:38.162860426 +0000 UTC m=+100.455889717" watchObservedRunningTime="2026-01-21 11:12:38.163189144 +0000 UTC m=+100.456218436" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.181046 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" podStartSLOduration=84.181034891 podStartE2EDuration="1m24.181034891s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:38.180655107 +0000 UTC m=+100.473684399" watchObservedRunningTime="2026-01-21 11:12:38.181034891 +0000 UTC m=+100.474064184" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.181210 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podStartSLOduration=84.181205713 podStartE2EDuration="1m24.181205713s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:38.170652772 +0000 UTC m=+100.463682064" watchObservedRunningTime="2026-01-21 11:12:38.181205713 +0000 UTC m=+100.474235005" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.258945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.258998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.259007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.259021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.259029 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.361290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.361339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.361350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.361368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.361379 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.463198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.463229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.463238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.463251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.463258 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.565360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.565411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.565421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.565438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.565450 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.666993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.667041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.667050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.667065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.667075 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.768879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.768922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.768932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.768946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.768969 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.870866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.870895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.870903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.870915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.870925 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.972907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.972939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.972948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.972988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.972998 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.049040 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.049046 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:39 crc kubenswrapper[4824]: E0121 11:12:39.049174 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:39 crc kubenswrapper[4824]: E0121 11:12:39.049254 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.052265 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 04:23:10.736867186 +0000 UTC Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.074922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.074952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.074978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.074992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.075003 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:39Z","lastTransitionTime":"2026-01-21T11:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.176903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.176935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.176943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.176973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.176989 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:39Z","lastTransitionTime":"2026-01-21T11:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.279003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.279071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.279082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.279106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.279120 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:39Z","lastTransitionTime":"2026-01-21T11:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.280662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.280691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.280700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.280711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.280720 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:39Z","lastTransitionTime":"2026-01-21T11:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.306346 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" podStartSLOduration=84.30632603 podStartE2EDuration="1m24.30632603s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:38.218257308 +0000 UTC m=+100.511286600" watchObservedRunningTime="2026-01-21 11:12:39.30632603 +0000 UTC m=+101.599355321" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.306905 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2"] Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.307292 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.308570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.308610 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.309520 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.309939 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.318073 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=85.318056528 podStartE2EDuration="1m25.318056528s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.317723341 +0000 UTC m=+101.610752633" watchObservedRunningTime="2026-01-21 11:12:39.318056528 +0000 UTC m=+101.611085820" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.327235 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=46.327217951 podStartE2EDuration="46.327217951s" podCreationTimestamp="2026-01-21 11:11:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.326564851 +0000 UTC m=+101.619594144" watchObservedRunningTime="2026-01-21 11:12:39.327217951 +0000 UTC m=+101.620247242" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.343589 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.343620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.343638 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.343666 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.343759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.344125 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-lv4rn" podStartSLOduration=85.344106837 podStartE2EDuration="1m25.344106837s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.344034691 +0000 UTC m=+101.637063983" watchObservedRunningTime="2026-01-21 11:12:39.344106837 +0000 UTC m=+101.637136129" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.352917 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-jr64s" podStartSLOduration=85.352901259 podStartE2EDuration="1m25.352901259s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.352323471 +0000 UTC m=+101.645352764" watchObservedRunningTime="2026-01-21 11:12:39.352901259 +0000 UTC m=+101.645930552" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.366075 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.366065917 podStartE2EDuration="25.366065917s" podCreationTimestamp="2026-01-21 11:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.365574161 +0000 UTC m=+101.658603463" watchObservedRunningTime="2026-01-21 11:12:39.366065917 +0000 UTC m=+101.659095209" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.375746 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=84.37572735 podStartE2EDuration="1m24.37572735s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.375543515 +0000 UTC m=+101.668572817" watchObservedRunningTime="2026-01-21 11:12:39.37572735 +0000 UTC m=+101.668756643" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.393837 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=84.393821465 podStartE2EDuration="1m24.393821465s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.391477915 +0000 UTC m=+101.684507207" watchObservedRunningTime="2026-01-21 11:12:39.393821465 +0000 UTC m=+101.686850757" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445461 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445514 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445572 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445697 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.446477 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.450441 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.462147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.617091 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.048352 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:40 crc kubenswrapper[4824]: E0121 11:12:40.048515 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.048536 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:40 crc kubenswrapper[4824]: E0121 11:12:40.048654 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.053037 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 00:15:47.311527941 +0000 UTC Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.053201 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.058353 4824 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.393461 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" event={"ID":"3f2fe1cd-26f3-4484-86c5-d71cda217f5f","Type":"ContainerStarted","Data":"ecaecae9b4699a601f5a9b20e08f49084d84f678317c3be0710aa3a5090e1411"} Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.393501 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" event={"ID":"3f2fe1cd-26f3-4484-86c5-d71cda217f5f","Type":"ContainerStarted","Data":"0208f03c73487747fdc2f6a321ec1667db4975ce87f6b1a134782fb4bc91eeb9"} Jan 21 11:12:41 crc kubenswrapper[4824]: I0121 11:12:41.048754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:41 crc kubenswrapper[4824]: E0121 11:12:41.048879 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:41 crc kubenswrapper[4824]: I0121 11:12:41.049495 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:41 crc kubenswrapper[4824]: E0121 11:12:41.049682 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:42 crc kubenswrapper[4824]: I0121 11:12:42.049127 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:42 crc kubenswrapper[4824]: I0121 11:12:42.049150 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:42 crc kubenswrapper[4824]: E0121 11:12:42.049274 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:42 crc kubenswrapper[4824]: E0121 11:12:42.049475 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:43 crc kubenswrapper[4824]: I0121 11:12:43.048904 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:43 crc kubenswrapper[4824]: I0121 11:12:43.048904 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:43 crc kubenswrapper[4824]: E0121 11:12:43.049016 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:43 crc kubenswrapper[4824]: E0121 11:12:43.049092 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:43 crc kubenswrapper[4824]: I0121 11:12:43.049598 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:43 crc kubenswrapper[4824]: E0121 11:12:43.049733 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:12:44 crc kubenswrapper[4824]: I0121 11:12:44.049109 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:44 crc kubenswrapper[4824]: I0121 11:12:44.049213 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:44 crc kubenswrapper[4824]: E0121 11:12:44.049361 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:44 crc kubenswrapper[4824]: E0121 11:12:44.049475 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:45 crc kubenswrapper[4824]: I0121 11:12:45.048596 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:45 crc kubenswrapper[4824]: I0121 11:12:45.048676 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:45 crc kubenswrapper[4824]: E0121 11:12:45.048716 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:45 crc kubenswrapper[4824]: E0121 11:12:45.048809 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:46 crc kubenswrapper[4824]: I0121 11:12:46.048716 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:46 crc kubenswrapper[4824]: I0121 11:12:46.048835 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:46 crc kubenswrapper[4824]: E0121 11:12:46.048975 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:46 crc kubenswrapper[4824]: E0121 11:12:46.049065 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:47 crc kubenswrapper[4824]: I0121 11:12:47.048843 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:47 crc kubenswrapper[4824]: I0121 11:12:47.048876 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:47 crc kubenswrapper[4824]: E0121 11:12:47.048994 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:47 crc kubenswrapper[4824]: E0121 11:12:47.049080 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:48 crc kubenswrapper[4824]: I0121 11:12:48.048507 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:48 crc kubenswrapper[4824]: I0121 11:12:48.048509 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:48 crc kubenswrapper[4824]: E0121 11:12:48.049423 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:48 crc kubenswrapper[4824]: E0121 11:12:48.049503 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.048704 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:49 crc kubenswrapper[4824]: E0121 11:12:49.048817 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.048709 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:49 crc kubenswrapper[4824]: E0121 11:12:49.049128 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.414163 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/1.log" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.414651 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/0.log" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.414683 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" containerID="49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f" exitCode=1 Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.414708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerDied","Data":"49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f"} Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.414744 4824 scope.go:117] "RemoveContainer" containerID="75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.415695 4824 scope.go:117] "RemoveContainer" containerID="49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f" Jan 21 11:12:49 crc kubenswrapper[4824]: E0121 11:12:49.416031 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-x94nb_openshift-multus(6b9588d1-5876-4c2c-8de5-e5c0ad46f04e)\"" pod="openshift-multus/multus-x94nb" podUID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.428422 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" podStartSLOduration=95.428394389 podStartE2EDuration="1m35.428394389s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:40.403405701 +0000 UTC m=+102.696434993" watchObservedRunningTime="2026-01-21 11:12:49.428394389 +0000 UTC m=+111.721423681" Jan 21 11:12:50 crc kubenswrapper[4824]: I0121 11:12:50.049240 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:50 crc kubenswrapper[4824]: I0121 11:12:50.049300 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:50 crc kubenswrapper[4824]: E0121 11:12:50.049373 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:50 crc kubenswrapper[4824]: E0121 11:12:50.049476 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:50 crc kubenswrapper[4824]: I0121 11:12:50.417729 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/1.log" Jan 21 11:12:51 crc kubenswrapper[4824]: I0121 11:12:51.048525 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:51 crc kubenswrapper[4824]: I0121 11:12:51.048570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:51 crc kubenswrapper[4824]: E0121 11:12:51.048655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:51 crc kubenswrapper[4824]: E0121 11:12:51.048733 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:52 crc kubenswrapper[4824]: I0121 11:12:52.048905 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:52 crc kubenswrapper[4824]: I0121 11:12:52.049056 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:52 crc kubenswrapper[4824]: E0121 11:12:52.049685 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:52 crc kubenswrapper[4824]: E0121 11:12:52.049821 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:53 crc kubenswrapper[4824]: I0121 11:12:53.049229 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:53 crc kubenswrapper[4824]: E0121 11:12:53.049360 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:53 crc kubenswrapper[4824]: I0121 11:12:53.049236 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:53 crc kubenswrapper[4824]: E0121 11:12:53.049545 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:54 crc kubenswrapper[4824]: I0121 11:12:54.049131 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:54 crc kubenswrapper[4824]: I0121 11:12:54.049182 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:54 crc kubenswrapper[4824]: E0121 11:12:54.049279 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:54 crc kubenswrapper[4824]: E0121 11:12:54.049416 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:55 crc kubenswrapper[4824]: I0121 11:12:55.048765 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:55 crc kubenswrapper[4824]: E0121 11:12:55.048860 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:55 crc kubenswrapper[4824]: I0121 11:12:55.048899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:55 crc kubenswrapper[4824]: E0121 11:12:55.049218 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.048992 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.049025 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:56 crc kubenswrapper[4824]: E0121 11:12:56.049112 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:56 crc kubenswrapper[4824]: E0121 11:12:56.049177 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.049630 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.432315 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/3.log" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.434246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.434587 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.641555 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podStartSLOduration=102.641537931 podStartE2EDuration="1m42.641537931s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:56.456586283 +0000 UTC m=+118.749615575" watchObservedRunningTime="2026-01-21 11:12:56.641537931 +0000 UTC m=+118.934567223" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.641934 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9rmnw"] Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.642049 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:56 crc kubenswrapper[4824]: E0121 11:12:56.642147 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:57 crc kubenswrapper[4824]: I0121 11:12:57.048569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:57 crc kubenswrapper[4824]: E0121 11:12:57.048860 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:57 crc kubenswrapper[4824]: I0121 11:12:57.048569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:57 crc kubenswrapper[4824]: E0121 11:12:57.048931 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:58 crc kubenswrapper[4824]: E0121 11:12:58.028734 4824 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 21 11:12:58 crc kubenswrapper[4824]: I0121 11:12:58.048558 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:58 crc kubenswrapper[4824]: E0121 11:12:58.050943 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:58 crc kubenswrapper[4824]: I0121 11:12:58.051009 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:58 crc kubenswrapper[4824]: E0121 11:12:58.051084 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:58 crc kubenswrapper[4824]: E0121 11:12:58.096949 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 21 11:12:59 crc kubenswrapper[4824]: I0121 11:12:59.048446 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:59 crc kubenswrapper[4824]: I0121 11:12:59.048495 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:59 crc kubenswrapper[4824]: E0121 11:12:59.048562 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:59 crc kubenswrapper[4824]: E0121 11:12:59.048700 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:13:00 crc kubenswrapper[4824]: I0121 11:13:00.049011 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:00 crc kubenswrapper[4824]: I0121 11:13:00.049015 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:00 crc kubenswrapper[4824]: E0121 11:13:00.049135 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:13:00 crc kubenswrapper[4824]: E0121 11:13:00.049327 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:13:00 crc kubenswrapper[4824]: I0121 11:13:00.049685 4824 scope.go:117] "RemoveContainer" containerID="49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f" Jan 21 11:13:00 crc kubenswrapper[4824]: I0121 11:13:00.446741 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/1.log" Jan 21 11:13:00 crc kubenswrapper[4824]: I0121 11:13:00.446793 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerStarted","Data":"80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470"} Jan 21 11:13:01 crc kubenswrapper[4824]: I0121 11:13:01.049224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:01 crc kubenswrapper[4824]: I0121 11:13:01.049257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:01 crc kubenswrapper[4824]: E0121 11:13:01.049365 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:13:01 crc kubenswrapper[4824]: E0121 11:13:01.049456 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:13:02 crc kubenswrapper[4824]: I0121 11:13:02.048933 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:02 crc kubenswrapper[4824]: I0121 11:13:02.048988 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:02 crc kubenswrapper[4824]: E0121 11:13:02.049120 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:13:02 crc kubenswrapper[4824]: E0121 11:13:02.049210 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:13:03 crc kubenswrapper[4824]: I0121 11:13:03.048405 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:03 crc kubenswrapper[4824]: I0121 11:13:03.048415 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:03 crc kubenswrapper[4824]: E0121 11:13:03.048523 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:13:03 crc kubenswrapper[4824]: E0121 11:13:03.048620 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.049090 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.049136 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.051449 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.051634 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.051709 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.052247 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 21 11:13:05 crc kubenswrapper[4824]: I0121 11:13:05.048200 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:05 crc kubenswrapper[4824]: I0121 11:13:05.048260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:05 crc kubenswrapper[4824]: I0121 11:13:05.049843 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 21 11:13:05 crc kubenswrapper[4824]: I0121 11:13:05.050009 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.709469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.731932 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-htg7q"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.732567 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.732731 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.733037 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.737816 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.737949 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738092 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738202 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738239 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738376 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738488 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738499 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738597 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7grz2"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738890 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739151 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739339 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739338 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739504 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739610 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739871 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739902 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.740289 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.740379 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s2n22"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.747225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.747744 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.747864 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.748029 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.748429 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.748739 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.749113 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.750185 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.750379 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.750548 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mh6ds"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.750828 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.750985 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.751108 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.751108 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.751545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.751827 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.752099 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.752505 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.753623 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754005 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754234 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754317 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754388 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754462 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754530 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754612 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754652 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754725 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754798 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754930 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.755080 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.755415 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f6c7b"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.755769 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.756326 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.756367 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.756640 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.757119 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.757635 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.767728 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.767837 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.780607 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.781174 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.781792 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.781948 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782167 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782307 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782456 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782589 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782704 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782823 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783033 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783171 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783348 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783450 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783558 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783706 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783849 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786192 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786480 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786504 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786758 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786818 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786930 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787013 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787103 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787262 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787505 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787608 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787967 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787985 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.788139 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.788162 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.788613 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790200 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790329 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790340 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790360 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790475 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790474 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790557 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790526 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790617 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790673 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790525 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790571 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790766 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796272 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796437 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796546 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796641 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796689 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796733 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796821 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.797268 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.797986 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.799468 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.799583 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8g4qk"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.800042 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.800670 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.801145 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.802362 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.802733 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.803230 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.803401 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.804183 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.806666 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.806886 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807231 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cpgg4"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807481 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807574 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807710 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807753 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807987 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808083 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808091 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-serving-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808109 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-trusted-ca-bundle\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808125 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-encryption-config\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808141 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zq9z\" (UniqueName: \"kubernetes.io/projected/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-kube-api-access-5zq9z\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808166 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49ec79bd-9eb8-42c8-a262-e96c1457f686-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808182 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg7z5\" (UniqueName: \"kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808196 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808222 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klfwh\" (UniqueName: \"kubernetes.io/projected/0475fd50-10a8-4195-8643-9408478e953c-kube-api-access-klfwh\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8qjw\" (UniqueName: \"kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808266 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808280 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-client\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808296 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-encryption-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808348 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh2qf\" (UniqueName: \"kubernetes.io/projected/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-kube-api-access-gh2qf\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808393 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-audit\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808431 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-serving-cert\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-config\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-config\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808480 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-service-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808517 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8d0d706a-5fe5-402d-9040-3c763e525554-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-trusted-ca\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-images\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808584 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808605 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-audit-dir\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808623 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808638 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skznz\" (UniqueName: \"kubernetes.io/projected/49ec79bd-9eb8-42c8-a262-e96c1457f686-kube-api-access-skznz\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808657 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-serving-cert\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d0d706a-5fe5-402d-9040-3c763e525554-serving-cert\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808689 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-node-pullsecrets\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808703 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808717 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl46h\" (UniqueName: \"kubernetes.io/projected/29d5a4fe-fd42-42ec-85b4-c79af6424296-kube-api-access-rl46h\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808746 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-audit-policies\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29d5a4fe-fd42-42ec-85b4-c79af6424296-serving-cert\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808780 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808794 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808808 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0475fd50-10a8-4195-8643-9408478e953c-audit-dir\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-config\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808863 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49ec79bd-9eb8-42c8-a262-e96c1457f686-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808891 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-image-import-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808905 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-etcd-client\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm7v7\" (UniqueName: \"kubernetes.io/projected/8364adda-37b5-4a9d-8abd-3daf12470f20-kube-api-access-fm7v7\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-serving-cert\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808946 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6nx5\" (UniqueName: \"kubernetes.io/projected/8d0d706a-5fe5-402d-9040-3c763e525554-kube-api-access-n6nx5\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808982 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.809248 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.809672 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.813317 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.813528 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.813653 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.815070 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.815479 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rdcbg"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.815693 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.815787 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.816256 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.816333 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.816744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.817599 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.818032 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.818189 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.818414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.818648 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.819042 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.820488 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.821331 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.821437 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-27d2q"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.822054 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.824174 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.824800 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.827622 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.835472 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.836538 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c5xks"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.837715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.838677 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.843087 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.844873 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.850886 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-bblkp"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.851948 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.852514 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.853185 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.853637 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.854310 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.854641 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.856304 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.856715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.857361 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7grz2"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.859187 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.860431 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.861417 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s2n22"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.863139 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.864336 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-htg7q"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.865191 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.865941 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dlgmw"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.866805 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-5s88g"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.867036 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.867341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.867860 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.868482 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.869481 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mh6ds"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.870541 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c5xks"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.871394 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.872246 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.873230 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.873732 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.874045 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.875034 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.876236 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.877289 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.878005 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.879644 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8g4qk"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.880488 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cpgg4"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.881327 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rdcbg"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.882149 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.883000 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.883797 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.884594 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.885494 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.886785 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.887665 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.888724 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.889592 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.890422 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.891266 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.892092 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-27d2q"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.892921 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f6c7b"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.896840 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jgvgl"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.897614 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.899687 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dlgmw"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.899712 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.899784 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.899981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jgvgl"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.910398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-serving-cert\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.910433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-config\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.910475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-config\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.910494 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-service-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8d0d706a-5fe5-402d-9040-3c763e525554-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911126 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-trusted-ca\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911151 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-images\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skznz\" (UniqueName: \"kubernetes.io/projected/49ec79bd-9eb8-42c8-a262-e96c1457f686-kube-api-access-skznz\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911226 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-audit-dir\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911243 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911265 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-serving-cert\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911284 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d0d706a-5fe5-402d-9040-3c763e525554-serving-cert\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-node-pullsecrets\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911321 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911348 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl46h\" (UniqueName: \"kubernetes.io/projected/29d5a4fe-fd42-42ec-85b4-c79af6424296-kube-api-access-rl46h\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911368 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-audit-policies\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29d5a4fe-fd42-42ec-85b4-c79af6424296-serving-cert\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911434 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-config\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911538 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911573 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0475fd50-10a8-4195-8643-9408478e953c-audit-dir\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-config\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911618 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49ec79bd-9eb8-42c8-a262-e96c1457f686-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-image-import-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-etcd-client\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911669 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8d0d706a-5fe5-402d-9040-3c763e525554-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm7v7\" (UniqueName: \"kubernetes.io/projected/8364adda-37b5-4a9d-8abd-3daf12470f20-kube-api-access-fm7v7\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911742 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-serving-cert\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911791 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6nx5\" (UniqueName: \"kubernetes.io/projected/8d0d706a-5fe5-402d-9040-3c763e525554-kube-api-access-n6nx5\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911813 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-serving-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911849 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911867 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49ec79bd-9eb8-42c8-a262-e96c1457f686-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-trusted-ca-bundle\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911897 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-encryption-config\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911911 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zq9z\" (UniqueName: \"kubernetes.io/projected/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-kube-api-access-5zq9z\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911930 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg7z5\" (UniqueName: \"kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911945 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911977 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klfwh\" (UniqueName: \"kubernetes.io/projected/0475fd50-10a8-4195-8643-9408478e953c-kube-api-access-klfwh\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8qjw\" (UniqueName: \"kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912014 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-node-pullsecrets\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912021 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-client\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912110 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-encryption-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912152 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh2qf\" (UniqueName: \"kubernetes.io/projected/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-kube-api-access-gh2qf\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-audit\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-audit-policies\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-config\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912755 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-trusted-ca\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-service-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.913194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-audit\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.913216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-audit-dir\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.913350 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-serving-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.913659 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-config\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.913941 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.914357 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.914571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.914652 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.914822 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49ec79bd-9eb8-42c8-a262-e96c1457f686-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.915484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-images\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.915695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-image-import-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.915730 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-serving-cert\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.915782 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0475fd50-10a8-4195-8643-9408478e953c-audit-dir\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.915826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d0d706a-5fe5-402d-9040-3c763e525554-serving-cert\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.916131 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.916383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.916432 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-trusted-ca-bundle\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.916895 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-encryption-config\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-client\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917650 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917660 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29d5a4fe-fd42-42ec-85b4-c79af6424296-serving-cert\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917757 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917782 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-etcd-client\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917884 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-serving-cert\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-serving-cert\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.918024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.918696 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49ec79bd-9eb8-42c8-a262-e96c1457f686-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.919428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.919760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.919894 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-encryption-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.932535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.945712 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-b7tsx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.946260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.951654 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b7tsx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.954010 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.974078 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.994134 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.014089 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.035409 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.054160 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.075114 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.095173 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.114216 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.134675 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.154078 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.174105 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.194883 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.214157 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.234647 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.254213 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.274700 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.294223 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.314797 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.334873 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.354472 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.374432 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.394649 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.414543 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.435818 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.454279 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.474891 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.504934 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.514313 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.533938 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.554633 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.574338 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.594684 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.614019 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.634642 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.653906 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.674712 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.699321 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.714210 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.734752 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.755830 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.774930 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.794438 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.814750 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.833819 4824 request.go:700] Waited for 1.012475708s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.834781 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.853861 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.874430 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.895060 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.914525 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.934131 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.954468 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.974319 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.993897 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.013941 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.034359 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.055232 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.073947 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.094645 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.114180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.134764 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.154045 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.175337 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.194033 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.214311 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.234830 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.254064 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.273931 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.294349 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.314547 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.334793 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.353916 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.374368 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.393877 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.434380 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.454681 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.474535 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.494196 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.514863 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.533717 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.554263 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.574361 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.594504 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.614299 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.634024 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.654227 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.674174 4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.694277 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.714530 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.746270 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skznz\" (UniqueName: \"kubernetes.io/projected/49ec79bd-9eb8-42c8-a262-e96c1457f686-kube-api-access-skznz\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.764473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm7v7\" (UniqueName: \"kubernetes.io/projected/8364adda-37b5-4a9d-8abd-3daf12470f20-kube-api-access-fm7v7\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.785821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl46h\" (UniqueName: \"kubernetes.io/projected/29d5a4fe-fd42-42ec-85b4-c79af6424296-kube-api-access-rl46h\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.805156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6nx5\" (UniqueName: \"kubernetes.io/projected/8d0d706a-5fe5-402d-9040-3c763e525554-kube-api-access-n6nx5\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.824426 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg7z5\" (UniqueName: \"kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.844862 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh2qf\" (UniqueName: \"kubernetes.io/projected/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-kube-api-access-gh2qf\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.853733 4824 request.go:700] Waited for 1.937689926s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-oauth-apiserver/serviceaccounts/oauth-apiserver-sa/token Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.853867 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.863359 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.869342 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klfwh\" (UniqueName: \"kubernetes.io/projected/0475fd50-10a8-4195-8643-9408478e953c-kube-api-access-klfwh\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.887743 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8qjw\" (UniqueName: \"kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.890316 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.908834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zq9z\" (UniqueName: \"kubernetes.io/projected/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-kube-api-access-5zq9z\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.925350 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.935070 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.936886 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.955328 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.955736 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.962686 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.968493 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.975206 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.002156 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-htg7q"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.002891 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.019875 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8364adda_37b5_4a9d_8abd_3daf12470f20.slice/crio-57ced44b4bd056825dfdc76b73c5ce663d44e6ce7c97b61af5207421cc94a24c WatchSource:0}: Error finding container 57ced44b4bd056825dfdc76b73c5ce663d44e6ce7c97b61af5207421cc94a24c: Status 404 returned error can't find the container with id 57ced44b4bd056825dfdc76b73c5ce663d44e6ce7c97b61af5207421cc94a24c Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.025661 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgr59\" (UniqueName: \"kubernetes.io/projected/024ed20d-cf91-4660-998b-3d06d3464607-kube-api-access-xgr59\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033597 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q27ql\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-kube-api-access-q27ql\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13490666-4efb-492b-be88-1c6fde3f378e-serving-cert\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033675 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6sxp\" (UniqueName: \"kubernetes.io/projected/16be0b45-eab8-4d0a-823b-5d877031dba5-kube-api-access-h6sxp\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033691 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh969\" (UniqueName: \"kubernetes.io/projected/74172695-d17f-4a47-b013-9e025e534916-kube-api-access-fh969\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033723 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ksj6\" (UniqueName: \"kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033756 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c873221f-c31d-411a-bbf5-c59e45b473a5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033773 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033839 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-service-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033869 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/86dc1658-d302-449f-9ae3-d7bc5e795c57-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033885 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-key\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c873221f-c31d-411a-bbf5-c59e45b473a5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-serving-cert\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034045 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-cabundle\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034090 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjjkx\" (UniqueName: \"kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034107 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b79b805a-7ab1-467f-90a5-ac7944cc564e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034121 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ad02bb-f06a-4619-87d1-112c78642603-config\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034137 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q9tc\" (UniqueName: \"kubernetes.io/projected/4b22a7aa-bcc0-4876-a374-a9943a6040bd-kube-api-access-6q9tc\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034186 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d2c136-5241-41da-a630-3d70273f1670-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034214 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2552x\" (UniqueName: \"kubernetes.io/projected/27b93806-10a5-46ba-8a81-20b05fd990c1-kube-api-access-2552x\") pod \"migrator-59844c95c7-j6bst\" (UID: \"27b93806-10a5-46ba-8a81-20b05fd990c1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034239 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034255 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qg69\" (UniqueName: \"kubernetes.io/projected/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-kube-api-access-6qg69\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034272 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-metrics-tls\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-metrics-tls\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034349 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqqww\" (UniqueName: \"kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pczg5\" (UniqueName: \"kubernetes.io/projected/86dc1658-d302-449f-9ae3-d7bc5e795c57-kube-api-access-pczg5\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034441 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034462 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-etcd-client\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035308 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74172695-d17f-4a47-b013-9e025e534916-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlp6l\" (UniqueName: \"kubernetes.io/projected/6a0a5723-a022-41f3-9222-c62cac06397e-kube-api-access-hlp6l\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzrmn\" (UniqueName: \"kubernetes.io/projected/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-kube-api-access-dzrmn\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035537 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-srv-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035555 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8xhg\" (UniqueName: \"kubernetes.io/projected/37d2c136-5241-41da-a630-3d70273f1670-kube-api-access-j8xhg\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.035758 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:12.535744472 +0000 UTC m=+134.828773765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035878 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07ad02bb-f06a-4619-87d1-112c78642603-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhtz6\" (UniqueName: \"kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035989 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d2c136-5241-41da-a630-3d70273f1670-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036374 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036430 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-config\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036471 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07ad02bb-f06a-4619-87d1-112c78642603-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036593 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-profile-collector-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036787 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-auth-proxy-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037016 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a0a5723-a022-41f3-9222-c62cac06397e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037298 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037335 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk6sw\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-kube-api-access-nk6sw\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037377 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74172695-d17f-4a47-b013-9e025e534916-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037421 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkwb6\" (UniqueName: \"kubernetes.io/projected/c117483a-fe5c-423b-995e-9c2e7f59566a-kube-api-access-tkwb6\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037538 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwmbq\" (UniqueName: \"kubernetes.io/projected/7a5d2d02-347e-4799-a28f-f303d51758d0-kube-api-access-pwmbq\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037554 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b22a7aa-bcc0-4876-a374-a9943a6040bd-service-ca-bundle\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037579 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13490666-4efb-492b-be88-1c6fde3f378e-config\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcznm\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c117483a-fe5c-423b-995e-9c2e7f59566a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038469 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-config\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038520 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-trusted-ca\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038538 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-metrics-certs\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c873221f-c31d-411a-bbf5-c59e45b473a5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b79b805a-7ab1-467f-90a5-ac7944cc564e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038604 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmsvd\" (UniqueName: \"kubernetes.io/projected/3358782c-3435-4c68-9f9c-c1040a6ada32-kube-api-access-pmsvd\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7a5d2d02-347e-4799-a28f-f303d51758d0-machine-approver-tls\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038722 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038745 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039081 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039103 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039143 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039157 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-default-certificate\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039186 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md94v\" (UniqueName: \"kubernetes.io/projected/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-kube-api-access-md94v\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039221 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsn9x\" (UniqueName: \"kubernetes.io/projected/2279132e-c274-424b-8c8b-80ae5bd7dfe5-kube-api-access-jsn9x\") pod \"downloads-7954f5f757-8g4qk\" (UID: \"2279132e-c274-424b-8c8b-80ae5bd7dfe5\") " pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039250 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-stats-auth\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039279 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-srv-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrlgj\" (UniqueName: \"kubernetes.io/projected/13490666-4efb-492b-be88-1c6fde3f378e-kube-api-access-qrlgj\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.061790 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s2n22"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140651 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140821 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74172695-d17f-4a47-b013-9e025e534916-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d93236c5-17ab-4cd0-88a3-f4b3479067e5-proxy-tls\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140877 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-config-volume\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140898 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlp6l\" (UniqueName: \"kubernetes.io/projected/6a0a5723-a022-41f3-9222-c62cac06397e-kube-api-access-hlp6l\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140916 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-srv-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8xhg\" (UniqueName: \"kubernetes.io/projected/37d2c136-5241-41da-a630-3d70273f1670-kube-api-access-j8xhg\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzrmn\" (UniqueName: \"kubernetes.io/projected/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-kube-api-access-dzrmn\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140992 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07ad02bb-f06a-4619-87d1-112c78642603-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141142 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d2c136-5241-41da-a630-3d70273f1670-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/faede743-203b-4bd1-b17a-776a4744bdf1-cert\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141173 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhtz6\" (UniqueName: \"kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j48kx\" (UniqueName: \"kubernetes.io/projected/23d7bb25-5d2c-4530-86dd-25a956953a4f-kube-api-access-j48kx\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141215 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141248 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-config\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141275 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-profile-collector-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141291 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07ad02bb-f06a-4619-87d1-112c78642603-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141322 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nqck\" (UniqueName: \"kubernetes.io/projected/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-kube-api-access-6nqck\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz5wq\" (UniqueName: \"kubernetes.io/projected/26739e4d-344c-4e25-bd0a-019f11456897-kube-api-access-nz5wq\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141380 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-auth-proxy-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141416 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a0a5723-a022-41f3-9222-c62cac06397e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141431 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-apiservice-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141458 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk6sw\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-kube-api-access-nk6sw\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74172695-d17f-4a47-b013-9e025e534916-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141505 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-csi-data-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141537 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkwb6\" (UniqueName: \"kubernetes.io/projected/c117483a-fe5c-423b-995e-9c2e7f59566a-kube-api-access-tkwb6\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwmbq\" (UniqueName: \"kubernetes.io/projected/7a5d2d02-347e-4799-a28f-f303d51758d0-kube-api-access-pwmbq\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141566 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b22a7aa-bcc0-4876-a374-a9943a6040bd-service-ca-bundle\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141624 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13490666-4efb-492b-be88-1c6fde3f378e-config\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-images\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5dpw\" (UniqueName: \"kubernetes.io/projected/faede743-203b-4bd1-b17a-776a4744bdf1-kube-api-access-h5dpw\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141672 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcznm\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141703 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63c09cd8-12ad-40a1-9877-3fc1500acb7a-proxy-tls\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c117483a-fe5c-423b-995e-9c2e7f59566a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxstx\" (UniqueName: \"kubernetes.io/projected/d93236c5-17ab-4cd0-88a3-f4b3479067e5-kube-api-access-kxstx\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141768 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-config\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141782 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-trusted-ca\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141796 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-metrics-certs\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141811 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d93236c5-17ab-4cd0-88a3-f4b3479067e5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141826 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b79b805a-7ab1-467f-90a5-ac7944cc564e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141846 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmsvd\" (UniqueName: \"kubernetes.io/projected/3358782c-3435-4c68-9f9c-c1040a6ada32-kube-api-access-pmsvd\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-metrics-tls\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c873221f-c31d-411a-bbf5-c59e45b473a5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141892 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7a5d2d02-347e-4799-a28f-f303d51758d0-machine-approver-tls\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141905 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141921 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141996 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk25w\" (UniqueName: \"kubernetes.io/projected/920ef22d-4a57-455c-a30f-c6d5f3a419e2-kube-api-access-lk25w\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142015 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142030 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-registration-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142046 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-default-certificate\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142060 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-webhook-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md94v\" (UniqueName: \"kubernetes.io/projected/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-kube-api-access-md94v\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142105 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsn9x\" (UniqueName: \"kubernetes.io/projected/2279132e-c274-424b-8c8b-80ae5bd7dfe5-kube-api-access-jsn9x\") pod \"downloads-7954f5f757-8g4qk\" (UID: \"2279132e-c274-424b-8c8b-80ae5bd7dfe5\") " pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142131 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-stats-auth\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-srv-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142160 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrlgj\" (UniqueName: \"kubernetes.io/projected/13490666-4efb-492b-be88-1c6fde3f378e-kube-api-access-qrlgj\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142175 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142191 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgr59\" (UniqueName: \"kubernetes.io/projected/024ed20d-cf91-4660-998b-3d06d3464607-kube-api-access-xgr59\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q27ql\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-kube-api-access-q27ql\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142261 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-plugins-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6sxp\" (UniqueName: \"kubernetes.io/projected/16be0b45-eab8-4d0a-823b-5d877031dba5-kube-api-access-h6sxp\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13490666-4efb-492b-be88-1c6fde3f378e-serving-cert\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ksj6\" (UniqueName: \"kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142323 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh969\" (UniqueName: \"kubernetes.io/projected/74172695-d17f-4a47-b013-9e025e534916-kube-api-access-fh969\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c873221f-c31d-411a-bbf5-c59e45b473a5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4hmd\" (UniqueName: \"kubernetes.io/projected/63c09cd8-12ad-40a1-9877-3fc1500acb7a-kube-api-access-s4hmd\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142397 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-service-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142411 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-socket-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142427 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/86dc1658-d302-449f-9ae3-d7bc5e795c57-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142442 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-key\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142456 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-node-bootstrap-token\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c873221f-c31d-411a-bbf5-c59e45b473a5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-mountpoint-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142502 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-serving-cert\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/23d7bb25-5d2c-4530-86dd-25a956953a4f-tmpfs\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142546 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-certs\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-cabundle\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjjkx\" (UniqueName: \"kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142600 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142621 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b79b805a-7ab1-467f-90a5-ac7944cc564e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142637 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142653 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d2c136-5241-41da-a630-3d70273f1670-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142683 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2552x\" (UniqueName: \"kubernetes.io/projected/27b93806-10a5-46ba-8a81-20b05fd990c1-kube-api-access-2552x\") pod \"migrator-59844c95c7-j6bst\" (UID: \"27b93806-10a5-46ba-8a81-20b05fd990c1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ad02bb-f06a-4619-87d1-112c78642603-config\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142712 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q9tc\" (UniqueName: \"kubernetes.io/projected/4b22a7aa-bcc0-4876-a374-a9943a6040bd-kube-api-access-6q9tc\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142734 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qg69\" (UniqueName: \"kubernetes.io/projected/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-kube-api-access-6qg69\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-metrics-tls\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142778 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-metrics-tls\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142809 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqqww\" (UniqueName: \"kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pczg5\" (UniqueName: \"kubernetes.io/projected/86dc1658-d302-449f-9ae3-d7bc5e795c57-kube-api-access-pczg5\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142879 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142908 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-etcd-client\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142923 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.143611 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.143680 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:12.643668409 +0000 UTC m=+134.936697701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.144181 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c873221f-c31d-411a-bbf5-c59e45b473a5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.144774 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.144938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-config\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.145168 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ad02bb-f06a-4619-87d1-112c78642603-config\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.146162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13490666-4efb-492b-be88-1c6fde3f378e-config\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.146410 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b22a7aa-bcc0-4876-a374-a9943a6040bd-service-ca-bundle\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.146579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.147131 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.147306 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.148487 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.148750 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-default-certificate\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.149072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.149601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.150817 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b79b805a-7ab1-467f-90a5-ac7944cc564e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.152791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153053 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153128 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-metrics-certs\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153214 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153456 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b79b805a-7ab1-467f-90a5-ac7944cc564e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153518 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d2c136-5241-41da-a630-3d70273f1670-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153884 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.154172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.154318 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74172695-d17f-4a47-b013-9e025e534916-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.154360 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-trusted-ca\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.154574 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.154972 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a0a5723-a022-41f3-9222-c62cac06397e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155046 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155668 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-srv-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155722 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7a5d2d02-347e-4799-a28f-f303d51758d0-machine-approver-tls\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155725 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-etcd-client\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156032 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-service-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156093 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156242 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07ad02bb-f06a-4619-87d1-112c78642603-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156345 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74172695-d17f-4a47-b013-9e025e534916-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156356 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-metrics-tls\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156864 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-cabundle\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.157517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-config\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.157735 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-auth-proxy-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-profile-collector-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158439 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158513 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-serving-cert\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158829 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158926 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159160 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d2c136-5241-41da-a630-3d70273f1670-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159478 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-stats-auth\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159928 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.160070 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13490666-4efb-492b-be88-1c6fde3f378e-serving-cert\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.160117 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-metrics-tls\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.160324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.160616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-srv-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.161423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-key\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.161984 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c873221f-c31d-411a-bbf5-c59e45b473a5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.164375 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c117483a-fe5c-423b-995e-9c2e7f59566a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.164760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/86dc1658-d302-449f-9ae3-d7bc5e795c57-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.168523 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.182629 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.186288 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk6sw\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-kube-api-access-nk6sw\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.186371 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0475fd50_10a8_4195_8643_9408478e953c.slice/crio-75c7fa16af4bca084c4cc4bde49251029592fd9bed3decc1ddef1b9a08515ea4 WatchSource:0}: Error finding container 75c7fa16af4bca084c4cc4bde49251029592fd9bed3decc1ddef1b9a08515ea4: Status 404 returned error can't find the container with id 75c7fa16af4bca084c4cc4bde49251029592fd9bed3decc1ddef1b9a08515ea4 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.206338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlp6l\" (UniqueName: \"kubernetes.io/projected/6a0a5723-a022-41f3-9222-c62cac06397e-kube-api-access-hlp6l\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.228171 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2552x\" (UniqueName: \"kubernetes.io/projected/27b93806-10a5-46ba-8a81-20b05fd990c1-kube-api-access-2552x\") pod \"migrator-59844c95c7-j6bst\" (UID: \"27b93806-10a5-46ba-8a81-20b05fd990c1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-mountpoint-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/23d7bb25-5d2c-4530-86dd-25a956953a4f-tmpfs\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-certs\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d93236c5-17ab-4cd0-88a3-f4b3479067e5-proxy-tls\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245331 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-config-volume\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245372 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/faede743-203b-4bd1-b17a-776a4744bdf1-cert\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245421 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j48kx\" (UniqueName: \"kubernetes.io/projected/23d7bb25-5d2c-4530-86dd-25a956953a4f-kube-api-access-j48kx\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245471 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nqck\" (UniqueName: \"kubernetes.io/projected/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-kube-api-access-6nqck\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245496 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz5wq\" (UniqueName: \"kubernetes.io/projected/26739e4d-344c-4e25-bd0a-019f11456897-kube-api-access-nz5wq\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-apiservice-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-csi-data-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245600 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-images\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245619 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5dpw\" (UniqueName: \"kubernetes.io/projected/faede743-203b-4bd1-b17a-776a4744bdf1-kube-api-access-h5dpw\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245641 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63c09cd8-12ad-40a1-9877-3fc1500acb7a-proxy-tls\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxstx\" (UniqueName: \"kubernetes.io/projected/d93236c5-17ab-4cd0-88a3-f4b3479067e5-kube-api-access-kxstx\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245682 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d93236c5-17ab-4cd0-88a3-f4b3479067e5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245712 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-metrics-tls\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245735 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk25w\" (UniqueName: \"kubernetes.io/projected/920ef22d-4a57-455c-a30f-c6d5f3a419e2-kube-api-access-lk25w\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245755 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-registration-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245777 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-webhook-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-plugins-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245917 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4hmd\" (UniqueName: \"kubernetes.io/projected/63c09cd8-12ad-40a1-9877-3fc1500acb7a-kube-api-access-s4hmd\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-socket-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245996 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-node-bootstrap-token\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.246091 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-csi-data-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.246424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-registration-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.246523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-mountpoint-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.247264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d93236c5-17ab-4cd0-88a3-f4b3479067e5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.247416 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/23d7bb25-5d2c-4530-86dd-25a956953a4f-tmpfs\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.247449 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.247497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-plugins-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.248341 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63c09cd8-12ad-40a1-9877-3fc1500acb7a-proxy-tls\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.248384 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-config-volume\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.248542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-socket-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.248671 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:12.748655901 +0000 UTC m=+135.041685193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.251527 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-images\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.253587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/faede743-203b-4bd1-b17a-776a4744bdf1-cert\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.255332 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-apiservice-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.255332 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-certs\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.255760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d93236c5-17ab-4cd0-88a3-f4b3479067e5-proxy-tls\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.255866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-node-bootstrap-token\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.256083 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqqww\" (UniqueName: \"kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.256176 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-metrics-tls\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.258467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-webhook-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.265531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pczg5\" (UniqueName: \"kubernetes.io/projected/86dc1658-d302-449f-9ae3-d7bc5e795c57-kube-api-access-pczg5\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.280218 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.285702 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkwb6\" (UniqueName: \"kubernetes.io/projected/c117483a-fe5c-423b-995e-9c2e7f59566a-kube-api-access-tkwb6\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.295377 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7grz2"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.308407 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwmbq\" (UniqueName: \"kubernetes.io/projected/7a5d2d02-347e-4799-a28f-f303d51758d0-kube-api-access-pwmbq\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.317900 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.318976 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.330761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.338142 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d0d706a_5fe5_402d_9040_3c763e525554.slice/crio-c63ddc91317112389bd667f708219c27c993346cd6fab5c172fde1b836cc4835 WatchSource:0}: Error finding container c63ddc91317112389bd667f708219c27c993346cd6fab5c172fde1b836cc4835: Status 404 returned error can't find the container with id c63ddc91317112389bd667f708219c27c993346cd6fab5c172fde1b836cc4835 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.347648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.348683 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:12.848648883 +0000 UTC m=+135.141678175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.350629 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzrmn\" (UniqueName: \"kubernetes.io/projected/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-kube-api-access-dzrmn\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.365544 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.372723 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcznm\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.373561 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mh6ds"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.374929 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.387918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.388978 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc585ef2f_3e29_4b81_918f_11b075006ca0.slice/crio-0b16adc0798e89f22e1034c180951bc349e7c3b376c2bea592b22d2474b2bc5e WatchSource:0}: Error finding container 0b16adc0798e89f22e1034c180951bc349e7c3b376c2bea592b22d2474b2bc5e: Status 404 returned error can't find the container with id 0b16adc0798e89f22e1034c180951bc349e7c3b376c2bea592b22d2474b2bc5e Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.390560 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.411068 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qg69\" (UniqueName: \"kubernetes.io/projected/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-kube-api-access-6qg69\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.415571 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.429187 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.431386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q9tc\" (UniqueName: \"kubernetes.io/projected/4b22a7aa-bcc0-4876-a374-a9943a6040bd-kube-api-access-6q9tc\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.434094 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.441397 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.448598 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjjkx\" (UniqueName: \"kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.449736 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.453178 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:12.953164449 +0000 UTC m=+135.246193741 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.459805 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.467802 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c873221f-c31d-411a-bbf5-c59e45b473a5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.470240 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.479922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nwlwd" event={"ID":"c585ef2f-3e29-4b81-918f-11b075006ca0","Type":"ContainerStarted","Data":"0b16adc0798e89f22e1034c180951bc349e7c3b376c2bea592b22d2474b2bc5e"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.487719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" event={"ID":"c4b9a738-6a66-467d-b8a5-bddb20cb7e98","Type":"ContainerStarted","Data":"a4408a690301e0b3fde8a526b30b6f1091bb7fa680fdd7aff2e9a023eba39e40"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.487750 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" event={"ID":"c4b9a738-6a66-467d-b8a5-bddb20cb7e98","Type":"ContainerStarted","Data":"94a67f23ca38cf50838a0b0ca9ef557570584d59359aa614c529d238934cbf0e"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.490324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.499426 4824 generic.go:334] "Generic (PLEG): container finished" podID="0475fd50-10a8-4195-8643-9408478e953c" containerID="8e7f98988a1c14c829902e8582cad543e1a6a6f7682b17a425252e44a603a94a" exitCode=0 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.499493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" event={"ID":"0475fd50-10a8-4195-8643-9408478e953c","Type":"ContainerDied","Data":"8e7f98988a1c14c829902e8582cad543e1a6a6f7682b17a425252e44a603a94a"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.499519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" event={"ID":"0475fd50-10a8-4195-8643-9408478e953c","Type":"ContainerStarted","Data":"75c7fa16af4bca084c4cc4bde49251029592fd9bed3decc1ddef1b9a08515ea4"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.504351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" event={"ID":"49ec79bd-9eb8-42c8-a262-e96c1457f686","Type":"ContainerStarted","Data":"24756c15affe3985216342ed2ac894d5b6bf6b2cb876e0f10ccb2500b9bad960"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.504388 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" event={"ID":"49ec79bd-9eb8-42c8-a262-e96c1457f686","Type":"ContainerStarted","Data":"b847215d00329a56ffa445c815cb09f06304f7917a8c399ea195f25ad640e745"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.506759 4824 generic.go:334] "Generic (PLEG): container finished" podID="8364adda-37b5-4a9d-8abd-3daf12470f20" containerID="01624aa0fea408599e8ca78243d73a26a5d305f3b193a4bc7fb78578ffb7fe1c" exitCode=0 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.510839 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" event={"ID":"8364adda-37b5-4a9d-8abd-3daf12470f20","Type":"ContainerDied","Data":"01624aa0fea408599e8ca78243d73a26a5d305f3b193a4bc7fb78578ffb7fe1c"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.510884 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" event={"ID":"8364adda-37b5-4a9d-8abd-3daf12470f20","Type":"ContainerStarted","Data":"57ced44b4bd056825dfdc76b73c5ce663d44e6ce7c97b61af5207421cc94a24c"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.515015 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsn9x\" (UniqueName: \"kubernetes.io/projected/2279132e-c274-424b-8c8b-80ae5bd7dfe5-kube-api-access-jsn9x\") pod \"downloads-7954f5f757-8g4qk\" (UID: \"2279132e-c274-424b-8c8b-80ae5bd7dfe5\") " pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.524505 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" event={"ID":"d962d0be-438a-4d08-8113-525238fa2156","Type":"ContainerStarted","Data":"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.524711 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" event={"ID":"d962d0be-438a-4d08-8113-525238fa2156","Type":"ContainerStarted","Data":"90b18225419fa36b91ae49d12b220ac91e071e5928e03c8bbf47f6f6ddb46af0"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.525018 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.525166 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b22a7aa_bcc0_4876_a374_a9943a6040bd.slice/crio-3d55b184fa201e03b48f73696a0f95c5e0b4f0887605e69e2620a010c15909bf WatchSource:0}: Error finding container 3d55b184fa201e03b48f73696a0f95c5e0b4f0887605e69e2620a010c15909bf: Status 404 returned error can't find the container with id 3d55b184fa201e03b48f73696a0f95c5e0b4f0887605e69e2620a010c15909bf Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.527089 4824 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4dkd8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.527146 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" podUID="d962d0be-438a-4d08-8113-525238fa2156" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.527286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" event={"ID":"8d0d706a-5fe5-402d-9040-3c763e525554","Type":"ContainerStarted","Data":"57c4a967c5b9aabd63ab037e2cb9dc29b3931b78fab3f197b33da1b36c5c1370"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.527311 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" event={"ID":"8d0d706a-5fe5-402d-9040-3c763e525554","Type":"ContainerStarted","Data":"c63ddc91317112389bd667f708219c27c993346cd6fab5c172fde1b836cc4835"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.534874 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" event={"ID":"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3","Type":"ContainerStarted","Data":"57b5bf1ffd0cee5939e39d5fe2a424668f24f6cb5ff7c71259477e24f21e95e6"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.534900 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" event={"ID":"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3","Type":"ContainerStarted","Data":"a7f32826f8477e3b840ca1f13a4b314f384df8cd972a12fc948ec057aef32d6f"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.538007 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" event={"ID":"29d5a4fe-fd42-42ec-85b4-c79af6424296","Type":"ContainerStarted","Data":"f0b3399b1909d22a75a5c79456700bfad80a9ed4c2b2ec54d4b775df25ba84ed"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.548399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8xhg\" (UniqueName: \"kubernetes.io/projected/37d2c136-5241-41da-a630-3d70273f1670-kube-api-access-j8xhg\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.550591 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.554023 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.054004091 +0000 UTC m=+135.347033383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.564208 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6sxp\" (UniqueName: \"kubernetes.io/projected/16be0b45-eab8-4d0a-823b-5d877031dba5-kube-api-access-h6sxp\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.570621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07ad02bb-f06a-4619-87d1-112c78642603-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.575727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.586318 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.589880 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.590664 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.602715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.606713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrlgj\" (UniqueName: \"kubernetes.io/projected/13490666-4efb-492b-be88-1c6fde3f378e-kube-api-access-qrlgj\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.615160 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.615472 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.636187 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md94v\" (UniqueName: \"kubernetes.io/projected/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-kube-api-access-md94v\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.640105 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.651305 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.652050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhtz6\" (UniqueName: \"kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.658870 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.661887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.662256 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.162243119 +0000 UTC m=+135.455272411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.664485 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86dc1658_d302_449f_9ae3_d7bc5e795c57.slice/crio-41bb1d013c63d1613e3f9736a883f4a3862526187c159bfbeb51c381ef8623d0 WatchSource:0}: Error finding container 41bb1d013c63d1613e3f9736a883f4a3862526187c159bfbeb51c381ef8623d0: Status 404 returned error can't find the container with id 41bb1d013c63d1613e3f9736a883f4a3862526187c159bfbeb51c381ef8623d0 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.666677 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh969\" (UniqueName: \"kubernetes.io/projected/74172695-d17f-4a47-b013-9e025e534916-kube-api-access-fh969\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.671553 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.686581 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.690387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmsvd\" (UniqueName: \"kubernetes.io/projected/3358782c-3435-4c68-9f9c-c1040a6ada32-kube-api-access-pmsvd\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.703933 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.707908 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q27ql\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-kube-api-access-q27ql\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.709329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.712277 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-27d2q"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.722513 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.729315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ksj6\" (UniqueName: \"kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.747063 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.749347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgr59\" (UniqueName: \"kubernetes.io/projected/024ed20d-cf91-4660-998b-3d06d3464607-kube-api-access-xgr59\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.755777 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.755827 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.759206 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.762917 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.763513 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.263496239 +0000 UTC m=+135.556525531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.767981 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7352d717_b2a3_4cd2_9e35_3ae2fe2d7b34.slice/crio-a58b1e840c1fbaaf4323bfd0d7baf951ec2faa2f74b8bef14349f6474ca6b7d8 WatchSource:0}: Error finding container a58b1e840c1fbaaf4323bfd0d7baf951ec2faa2f74b8bef14349f6474ca6b7d8: Status 404 returned error can't find the container with id a58b1e840c1fbaaf4323bfd0d7baf951ec2faa2f74b8bef14349f6474ca6b7d8 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.814635 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxstx\" (UniqueName: \"kubernetes.io/projected/d93236c5-17ab-4cd0-88a3-f4b3479067e5-kube-api-access-kxstx\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.840506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5dpw\" (UniqueName: \"kubernetes.io/projected/faede743-203b-4bd1-b17a-776a4744bdf1-kube-api-access-h5dpw\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.840810 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nqck\" (UniqueName: \"kubernetes.io/projected/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-kube-api-access-6nqck\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.859230 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz5wq\" (UniqueName: \"kubernetes.io/projected/26739e4d-344c-4e25-bd0a-019f11456897-kube-api-access-nz5wq\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.877697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.877823 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27b93806_10a5_46ba_8a81_20b05fd990c1.slice/crio-b90145e68496b029877d20e05a0410ab89139f34372c1cd90d9a9eea4f9469d9 WatchSource:0}: Error finding container b90145e68496b029877d20e05a0410ab89139f34372c1cd90d9a9eea4f9469d9: Status 404 returned error can't find the container with id b90145e68496b029877d20e05a0410ab89139f34372c1cd90d9a9eea4f9469d9 Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.878186 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.378171875 +0000 UTC m=+135.671201156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.885560 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.895633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk25w\" (UniqueName: \"kubernetes.io/projected/920ef22d-4a57-455c-a30f-c6d5f3a419e2-kube-api-access-lk25w\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.917622 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4hmd\" (UniqueName: \"kubernetes.io/projected/63c09cd8-12ad-40a1-9877-3fc1500acb7a-kube-api-access-s4hmd\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.920315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j48kx\" (UniqueName: \"kubernetes.io/projected/23d7bb25-5d2c-4530-86dd-25a956953a4f-kube-api-access-j48kx\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.945003 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.978344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.978727 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.478713617 +0000 UTC m=+135.771742909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.978823 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.997057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.045943 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.066862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.072642 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.083442 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.083974 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.084293 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.584278313 +0000 UTC m=+135.877307605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.095858 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.096218 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.110037 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.116142 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.184877 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.184978 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn"] Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.185232 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.685218384 +0000 UTC m=+135.978247676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: W0121 11:13:13.220028 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16be0b45_eab8_4d0a_823b_5d877031dba5.slice/crio-5fa3fc155f4cf8db417ce9eee59a7f91effb645e0b1ba49d021da1f9ce053d7f WatchSource:0}: Error finding container 5fa3fc155f4cf8db417ce9eee59a7f91effb645e0b1ba49d021da1f9ce053d7f: Status 404 returned error can't find the container with id 5fa3fc155f4cf8db417ce9eee59a7f91effb645e0b1ba49d021da1f9ce053d7f Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.288807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.289206 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.789182985 +0000 UTC m=+136.082212277 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.344714 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.391828 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.393170 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.893137317 +0000 UTC m=+136.186166609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.403170 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.403555 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.903542186 +0000 UTC m=+136.196571479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.444970 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f6c7b"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.455945 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.463111 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.463146 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.465721 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c5xks"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.506243 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.506490 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.006470431 +0000 UTC m=+136.299499723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.506633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.506887 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.006875271 +0000 UTC m=+136.299904564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.533383 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8g4qk"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.564547 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" event={"ID":"6a0a5723-a022-41f3-9222-c62cac06397e","Type":"ContainerStarted","Data":"a1cf9f71f9e51a9e84b32e1eb48baecb69e1af53943f073eac248f4bbd6e3490"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.564624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" event={"ID":"6a0a5723-a022-41f3-9222-c62cac06397e","Type":"ContainerStarted","Data":"24154a5c62917e47b8c541d57c970b6a4e0616e7f66b0df654db39d08754f71e"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.581279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" event={"ID":"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8","Type":"ContainerStarted","Data":"c0d9b7b7dc305ec1f4f68cddbad0ca766683d3701f7d30b811f77e3b1b850c5c"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.604244 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" event={"ID":"86dc1658-d302-449f-9ae3-d7bc5e795c57","Type":"ContainerStarted","Data":"863b0c3992a3892d462af79ff6fb9a86a7cdd408feecc5165baae4f4d4acded6"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.604296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" event={"ID":"86dc1658-d302-449f-9ae3-d7bc5e795c57","Type":"ContainerStarted","Data":"41bb1d013c63d1613e3f9736a883f4a3862526187c159bfbeb51c381ef8623d0"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.607378 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.607951 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.107934527 +0000 UTC m=+136.400963818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.609863 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" event={"ID":"27b93806-10a5-46ba-8a81-20b05fd990c1","Type":"ContainerStarted","Data":"b90145e68496b029877d20e05a0410ab89139f34372c1cd90d9a9eea4f9469d9"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.617585 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" event={"ID":"0475fd50-10a8-4195-8643-9408478e953c","Type":"ContainerStarted","Data":"eba6f1475ef962456ec8d4aca6659cffa1eee13b1d1aef2f9c235c7f6532e8c1"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.629650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" event={"ID":"16be0b45-eab8-4d0a-823b-5d877031dba5","Type":"ContainerStarted","Data":"5fa3fc155f4cf8db417ce9eee59a7f91effb645e0b1ba49d021da1f9ce053d7f"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.689634 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" event={"ID":"09bd0b9c-46fb-433f-bd30-014c69c90d9f","Type":"ContainerStarted","Data":"277011d8fed4b1333d53f5a4a084181f52dada1e6d85b31750d9bab9937b3678"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.689676 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" event={"ID":"09bd0b9c-46fb-433f-bd30-014c69c90d9f","Type":"ContainerStarted","Data":"68b9bb179c8bbe2beb17e913adc7df65d063a4b5a6cdcd18f32da44f14fdb4ae"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.719646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.720573 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.220562928 +0000 UTC m=+136.513592220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.723622 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" event={"ID":"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3","Type":"ContainerStarted","Data":"13825c9a9e1496a36936513a845471c8cd2d12f3e5dd6f1ea597f62b7a5881e7"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.778260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nwlwd" event={"ID":"c585ef2f-3e29-4b81-918f-11b075006ca0","Type":"ContainerStarted","Data":"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.785138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bblkp" event={"ID":"4b22a7aa-bcc0-4876-a374-a9943a6040bd","Type":"ContainerStarted","Data":"60ddd7b0675ec58022e55c78cbe86c12e0d7acd9251135287d12b3098481d22d"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.785175 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bblkp" event={"ID":"4b22a7aa-bcc0-4876-a374-a9943a6040bd","Type":"ContainerStarted","Data":"3d55b184fa201e03b48f73696a0f95c5e0b4f0887605e69e2620a010c15909bf"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.791584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" event={"ID":"d132606d-b1ae-4cb7-865e-e1065c2eb4a9","Type":"ContainerStarted","Data":"50860c3a46aa9f38581345e1681b00dcccbfdd69caedca48718ba3de06ce2862"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.791630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" event={"ID":"d132606d-b1ae-4cb7-865e-e1065c2eb4a9","Type":"ContainerStarted","Data":"d15d5576fb1d7a7584b073ba478a594a4c953db85611b5679f8498b6ed394016"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.803573 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" event={"ID":"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34","Type":"ContainerStarted","Data":"a58b1e840c1fbaaf4323bfd0d7baf951ec2faa2f74b8bef14349f6474ca6b7d8"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.805090 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5s88g" event={"ID":"920ef22d-4a57-455c-a30f-c6d5f3a419e2","Type":"ContainerStarted","Data":"4c429d9e761441d396247bc1947df1a4fc0f76fe16e4927368fc3f1b799adee0"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.816145 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" event={"ID":"8364adda-37b5-4a9d-8abd-3daf12470f20","Type":"ContainerStarted","Data":"85188d995a9d3025fb63f5bb716e23bca5109219b40e8522015af40804bea5d3"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.820636 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.821914 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.321900004 +0000 UTC m=+136.614929296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.827399 4824 generic.go:334] "Generic (PLEG): container finished" podID="8d0d706a-5fe5-402d-9040-3c763e525554" containerID="57c4a967c5b9aabd63ab037e2cb9dc29b3931b78fab3f197b33da1b36c5c1370" exitCode=0 Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.827667 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" event={"ID":"8d0d706a-5fe5-402d-9040-3c763e525554","Type":"ContainerDied","Data":"57c4a967c5b9aabd63ab037e2cb9dc29b3931b78fab3f197b33da1b36c5c1370"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.832050 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" event={"ID":"29d5a4fe-fd42-42ec-85b4-c79af6424296","Type":"ContainerStarted","Data":"42f8e34d6093d1ac6aa986143cc096ebe737677ee356032a8cfa97ce107372d0"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.833585 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.838470 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" event={"ID":"c117483a-fe5c-423b-995e-9c2e7f59566a","Type":"ContainerStarted","Data":"05346edfb2dc22987a7065666d0c7ed1f8a5cfaee344800819a0a26c4db5b9ae"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.843278 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" event={"ID":"07ad02bb-f06a-4619-87d1-112c78642603","Type":"ContainerStarted","Data":"7f2a98a80d971de8e219b03ef193886a6ae7e63f2383d5c3777335050c9f8efd"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.882350 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" event={"ID":"7a5d2d02-347e-4799-a28f-f303d51758d0","Type":"ContainerStarted","Data":"33f8d434af3d7825f96a18549c7957930822a3b9862435f1df2b941c88dc50b6"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.882386 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" event={"ID":"7a5d2d02-347e-4799-a28f-f303d51758d0","Type":"ContainerStarted","Data":"7208fa8eaa4127528fdc27fa8019ce379ca5e5017e76eebb7bbfc1571ac29f3d"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.898574 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.924946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.929253 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.429238932 +0000 UTC m=+136.722268224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.030385 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.030760 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.530745507 +0000 UTC m=+136.823774798 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.032062 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.052262 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.552244211 +0000 UTC m=+136.845273503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.094141 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" podStartSLOduration=120.094127845 podStartE2EDuration="2m0.094127845s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.094088532 +0000 UTC m=+136.387117824" watchObservedRunningTime="2026-01-21 11:13:14.094127845 +0000 UTC m=+136.387157138" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.134301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.134730 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.634713565 +0000 UTC m=+136.927742857 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.219707 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" podStartSLOduration=120.21969199 podStartE2EDuration="2m0.21969199s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.219005701 +0000 UTC m=+136.512034993" watchObservedRunningTime="2026-01-21 11:13:14.21969199 +0000 UTC m=+136.512721282" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.241718 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.242245 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.74223236 +0000 UTC m=+137.035261652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.297063 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.341370 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.342292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.342512 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.84249544 +0000 UTC m=+137.135524731 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.342543 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.342906 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.842898847 +0000 UTC m=+137.135928139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.362276 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.388205 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.392173 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.417935 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.445507 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.446546 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.946527247 +0000 UTC m=+137.239556538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.462042 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.469150 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:14 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:14 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:14 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.469214 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:14 crc kubenswrapper[4824]: W0121 11:13:14.478719 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51e27630_d5c1_4eed_8232_3698ab99fc19.slice/crio-66caec7ee49f045aa9a142ffed3e9d86fa441cb0eec919f264b1bbebed641a83 WatchSource:0}: Error finding container 66caec7ee49f045aa9a142ffed3e9d86fa441cb0eec919f264b1bbebed641a83: Status 404 returned error can't find the container with id 66caec7ee49f045aa9a142ffed3e9d86fa441cb0eec919f264b1bbebed641a83 Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.519643 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" podStartSLOduration=120.519627486 podStartE2EDuration="2m0.519627486s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.519052917 +0000 UTC m=+136.812082209" watchObservedRunningTime="2026-01-21 11:13:14.519627486 +0000 UTC m=+136.812656777" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.547696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.548051 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.048039412 +0000 UTC m=+137.341068705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.618998 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-nwlwd" podStartSLOduration=120.618951882 podStartE2EDuration="2m0.618951882s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.618397692 +0000 UTC m=+136.911426983" watchObservedRunningTime="2026-01-21 11:13:14.618951882 +0000 UTC m=+136.911981174" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.629517 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b7tsx"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.642703 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rdcbg"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.648999 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.649213 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.149190049 +0000 UTC m=+137.442219341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.649449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.649821 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.149806196 +0000 UTC m=+137.442835487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.655495 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" podStartSLOduration=120.655483347 podStartE2EDuration="2m0.655483347s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.652940402 +0000 UTC m=+136.945969693" watchObservedRunningTime="2026-01-21 11:13:14.655483347 +0000 UTC m=+136.948512639" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.703367 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.705037 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cpgg4"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.708736 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dlgmw"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.722024 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.722798 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-bblkp" podStartSLOduration=119.72278645 podStartE2EDuration="1m59.72278645s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.70327106 +0000 UTC m=+136.996300352" watchObservedRunningTime="2026-01-21 11:13:14.72278645 +0000 UTC m=+137.015815741" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.750882 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.751050 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.251027164 +0000 UTC m=+137.544056456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.751285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.751638 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.251623745 +0000 UTC m=+137.544653037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.805718 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" podStartSLOduration=119.805700658 podStartE2EDuration="1m59.805700658s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.770350011 +0000 UTC m=+137.063379303" watchObservedRunningTime="2026-01-21 11:13:14.805700658 +0000 UTC m=+137.098729951" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.806130 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" podStartSLOduration=119.806122791 podStartE2EDuration="1m59.806122791s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.804796621 +0000 UTC m=+137.097825913" watchObservedRunningTime="2026-01-21 11:13:14.806122791 +0000 UTC m=+137.099152083" Jan 21 11:13:14 crc kubenswrapper[4824]: W0121 11:13:14.813586 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23d7bb25_5d2c_4530_86dd_25a956953a4f.slice/crio-49d3a818b492b8df859fbc6c1e8a1a805dbadec0d6a740e52d27b04d4158d925 WatchSource:0}: Error finding container 49d3a818b492b8df859fbc6c1e8a1a805dbadec0d6a740e52d27b04d4158d925: Status 404 returned error can't find the container with id 49d3a818b492b8df859fbc6c1e8a1a805dbadec0d6a740e52d27b04d4158d925 Jan 21 11:13:14 crc kubenswrapper[4824]: W0121 11:13:14.834984 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7fe95fc_b7d1_4ad1_8acf_e7ee4564babe.slice/crio-a92e9a2d4247fb90b0a505fd7792042cdffd7c5469973ffe9021e3c2156f63eb WatchSource:0}: Error finding container a92e9a2d4247fb90b0a505fd7792042cdffd7c5469973ffe9021e3c2156f63eb: Status 404 returned error can't find the container with id a92e9a2d4247fb90b0a505fd7792042cdffd7c5469973ffe9021e3c2156f63eb Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.853683 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.854116 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.354104269 +0000 UTC m=+137.647133561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: W0121 11:13:14.872233 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74172695_d17f_4a47_b013_9e025e534916.slice/crio-2bb9c7e7f018544ad416d59b3c538a5f29156538646e38a8327e89c2bd11cd88 WatchSource:0}: Error finding container 2bb9c7e7f018544ad416d59b3c538a5f29156538646e38a8327e89c2bd11cd88: Status 404 returned error can't find the container with id 2bb9c7e7f018544ad416d59b3c538a5f29156538646e38a8327e89c2bd11cd88 Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.893553 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" event={"ID":"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8","Type":"ContainerStarted","Data":"757f71987259387576ae95787f25d067b1c7a3e6c5dcd0206a1faa0653381151"} Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.895707 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.915798 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-j8dcx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.915836 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" podUID="02c3b11b-dbc4-445d-94fa-95c5ccc03cb8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.916448 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" event={"ID":"74172695-d17f-4a47-b013-9e025e534916","Type":"ContainerStarted","Data":"2bb9c7e7f018544ad416d59b3c538a5f29156538646e38a8327e89c2bd11cd88"} Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.921625 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" podStartSLOduration=119.921613037 podStartE2EDuration="1m59.921613037s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.854312058 +0000 UTC m=+137.147341351" watchObservedRunningTime="2026-01-21 11:13:14.921613037 +0000 UTC m=+137.214642328" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.925146 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.934878 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jgvgl"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.947421 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" event={"ID":"16be0b45-eab8-4d0a-823b-5d877031dba5","Type":"ContainerStarted","Data":"dd377815656d0707d5d6e9d707e7afef4665367c1db4361ef545e1a8bf287d97"} Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.948614 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.959092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.959365 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.459354183 +0000 UTC m=+137.752383475 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.991206 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" podStartSLOduration=119.991188828 podStartE2EDuration="1m59.991188828s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.991022154 +0000 UTC m=+137.284051446" watchObservedRunningTime="2026-01-21 11:13:14.991188828 +0000 UTC m=+137.284218120" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:14.995393 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.000728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" event={"ID":"3358782c-3435-4c68-9f9c-c1040a6ada32","Type":"ContainerStarted","Data":"38a77399b318a83fe4ece3efc7fadb9b3a89f0203b4293698f30d603fa103891"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.010594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" event={"ID":"37d2c136-5241-41da-a630-3d70273f1670","Type":"ContainerStarted","Data":"7b51fb3348172ec85dd20f861b1f1bf51a4054b0bd408215322d12c361203133"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.010627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" event={"ID":"37d2c136-5241-41da-a630-3d70273f1670","Type":"ContainerStarted","Data":"923df253da37dd5b6251f3ca08df8620e565dc4afda0554844090400996c31f6"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.066540 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.067622 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.567608651 +0000 UTC m=+137.860637942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.093092 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" event={"ID":"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b","Type":"ContainerStarted","Data":"573d45c88ddfeee5cecbfabe535ab387ebbe06773fb6b02c5cc55e6a1d0fffa1"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.101012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" event={"ID":"86b5dca2-989f-4b6b-983b-87763ce0d469","Type":"ContainerStarted","Data":"046a7d0640fbf37bae3b158d9284d146391f1ce208f167ac80fea14196ea051f"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.102887 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" podStartSLOduration=120.102873276 podStartE2EDuration="2m0.102873276s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.090574559 +0000 UTC m=+137.383603851" watchObservedRunningTime="2026-01-21 11:13:15.102873276 +0000 UTC m=+137.395902567" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.132946 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" event={"ID":"13490666-4efb-492b-be88-1c6fde3f378e","Type":"ContainerStarted","Data":"cf24e15effeb0a5a7bbfe0c5b7d94b4824adff54c21b171c42502c9ba19f9242"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.133006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" event={"ID":"13490666-4efb-492b-be88-1c6fde3f378e","Type":"ContainerStarted","Data":"b1b1ae1f7226d453f6259d47fde178296c1e39b0d8660341bc42a84e5f5fc305"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.150368 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" podStartSLOduration=120.150351498 podStartE2EDuration="2m0.150351498s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.132179361 +0000 UTC m=+137.425208653" watchObservedRunningTime="2026-01-21 11:13:15.150351498 +0000 UTC m=+137.443380790" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.150739 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" podStartSLOduration=121.150734577 podStartE2EDuration="2m1.150734577s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.149610967 +0000 UTC m=+137.442640259" watchObservedRunningTime="2026-01-21 11:13:15.150734577 +0000 UTC m=+137.443763870" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.161789 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" event={"ID":"024ed20d-cf91-4660-998b-3d06d3464607","Type":"ContainerStarted","Data":"907f0e9f1ba0abbd21d3b2ebb8e7d365907202e44d50409f58a962b67e9c0f82"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.168323 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.169320 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.669310381 +0000 UTC m=+137.962339674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.187873 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" podStartSLOduration=120.18786142 podStartE2EDuration="2m0.18786142s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.187273095 +0000 UTC m=+137.480302386" watchObservedRunningTime="2026-01-21 11:13:15.18786142 +0000 UTC m=+137.480890712" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.196330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5s88g" event={"ID":"920ef22d-4a57-455c-a30f-c6d5f3a419e2","Type":"ContainerStarted","Data":"3b506a482bf40c00e2d6a4eeadde103dc948a6bbad06449b4b5cc81ed27d731f"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.244527 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-5s88g" podStartSLOduration=6.244510774 podStartE2EDuration="6.244510774s" podCreationTimestamp="2026-01-21 11:13:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.243365555 +0000 UTC m=+137.536394846" watchObservedRunningTime="2026-01-21 11:13:15.244510774 +0000 UTC m=+137.537540056" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.245450 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" podStartSLOduration=120.245436534 podStartE2EDuration="2m0.245436534s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.222323818 +0000 UTC m=+137.515353111" watchObservedRunningTime="2026-01-21 11:13:15.245436534 +0000 UTC m=+137.538465826" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.272493 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.273492 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.773475198 +0000 UTC m=+138.066504490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.286023 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" event={"ID":"51e27630-d5c1-4eed-8232-3698ab99fc19","Type":"ContainerStarted","Data":"66caec7ee49f045aa9a142ffed3e9d86fa441cb0eec919f264b1bbebed641a83"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.286194 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.294275 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" event={"ID":"27b93806-10a5-46ba-8a81-20b05fd990c1","Type":"ContainerStarted","Data":"66f680a3a461408efa0f9f7f82f5fd91d998ef92fe166391b9a76fcfcd379415"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.294339 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" event={"ID":"27b93806-10a5-46ba-8a81-20b05fd990c1","Type":"ContainerStarted","Data":"ac68dbd34db8c9a005fcd37771a0552c714428d27022925e4208aa9c867c7924"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.303541 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" podStartSLOduration=120.303529319 podStartE2EDuration="2m0.303529319s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.303186645 +0000 UTC m=+137.596215937" watchObservedRunningTime="2026-01-21 11:13:15.303529319 +0000 UTC m=+137.596558611" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.304324 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dlgmw" event={"ID":"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe","Type":"ContainerStarted","Data":"a92e9a2d4247fb90b0a505fd7792042cdffd7c5469973ffe9021e3c2156f63eb"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.309460 4824 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-qs8d6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.309521 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.325766 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8g4qk" event={"ID":"2279132e-c274-424b-8c8b-80ae5bd7dfe5","Type":"ContainerStarted","Data":"becc6485ca48ad42d37b3ff35bcbe4b1895daea3fb4e50341e5518aa753d7f77"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.326669 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.328117 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-8g4qk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.328157 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8g4qk" podUID="2279132e-c274-424b-8c8b-80ae5bd7dfe5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.338647 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" podStartSLOduration=120.338634265 podStartE2EDuration="2m0.338634265s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.335307187 +0000 UTC m=+137.628336479" watchObservedRunningTime="2026-01-21 11:13:15.338634265 +0000 UTC m=+137.631663556" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.366014 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" event={"ID":"b79b805a-7ab1-467f-90a5-ac7944cc564e","Type":"ContainerStarted","Data":"ff6b12f5197bc273b9b1f5eaa9fe8891d095a52df4dba219ea169c425cdbcc29"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.374079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.374790 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" event={"ID":"c117483a-fe5c-423b-995e-9c2e7f59566a","Type":"ContainerStarted","Data":"0afb8e9f166ddbd00968e6ab571144882d40df74bd23f93bf7a9e95410697ecc"} Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.377300 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.877280781 +0000 UTC m=+138.170310073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.386389 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8g4qk" podStartSLOduration=121.386370471 podStartE2EDuration="2m1.386370471s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.371765734 +0000 UTC m=+137.664795026" watchObservedRunningTime="2026-01-21 11:13:15.386370471 +0000 UTC m=+137.679399763" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.387033 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" event={"ID":"07ad02bb-f06a-4619-87d1-112c78642603","Type":"ContainerStarted","Data":"2d7963c8981edc983480fac716eac85cd99025b69e6ddc50c3242d01f5c53f13"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.399943 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" podStartSLOduration=120.399925456 podStartE2EDuration="2m0.399925456s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.399526557 +0000 UTC m=+137.692555849" watchObservedRunningTime="2026-01-21 11:13:15.399925456 +0000 UTC m=+137.692954749" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.402225 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b7tsx" event={"ID":"faede743-203b-4bd1-b17a-776a4744bdf1","Type":"ContainerStarted","Data":"409842eba4ded972cdf6e0d8b0cc3dfc9e94886c2d221b1cd5479b50e36f942f"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.446970 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" event={"ID":"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34","Type":"ContainerStarted","Data":"5c55044ce64ff381aa81be5bb4c9b389c84a5b9b921435cd1eeedf44d2c77613"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.447027 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" event={"ID":"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34","Type":"ContainerStarted","Data":"061b951668ab926a35a6e5500308fd9f83880f1a4fec23f785997d53ab76d390"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.466194 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:15 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:15 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:15 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.466265 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.467283 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" podStartSLOduration=120.467264386 podStartE2EDuration="2m0.467264386s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.425931645 +0000 UTC m=+137.718960937" watchObservedRunningTime="2026-01-21 11:13:15.467264386 +0000 UTC m=+137.760293678" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.467685 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" podStartSLOduration=120.467674637 podStartE2EDuration="2m0.467674637s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.466785627 +0000 UTC m=+137.759814919" watchObservedRunningTime="2026-01-21 11:13:15.467674637 +0000 UTC m=+137.760703928" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.478907 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.479295 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.979277747 +0000 UTC m=+138.272307039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.479421 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" event={"ID":"86dc1658-d302-449f-9ae3-d7bc5e795c57","Type":"ContainerStarted","Data":"ec7097ec47b1ede21a87b07e962cafae7130cdbdb00e1277f2a250754e3a5cd3"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.479848 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.480731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.482831 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.982811092 +0000 UTC m=+138.275840384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.488031 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" event={"ID":"14d99a1f-fca3-421a-8b47-ab4cda2f77fc","Type":"ContainerStarted","Data":"605262189eb74a464efb5cb3518becda7f59cb81cb5a3ce736fbc1a29ea45dc4"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.488088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" event={"ID":"14d99a1f-fca3-421a-8b47-ab4cda2f77fc","Type":"ContainerStarted","Data":"3217737690e95d2837429663404e46050468bcb19f6fbf887ccc71adb8493eef"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.488891 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.505047 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gjhp5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.505084 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.506762 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" podStartSLOduration=120.506742224 podStartE2EDuration="2m0.506742224s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.505342165 +0000 UTC m=+137.798371457" watchObservedRunningTime="2026-01-21 11:13:15.506742224 +0000 UTC m=+137.799771516" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.511669 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" event={"ID":"23d7bb25-5d2c-4530-86dd-25a956953a4f","Type":"ContainerStarted","Data":"49d3a818b492b8df859fbc6c1e8a1a805dbadec0d6a740e52d27b04d4158d925"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.542337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" event={"ID":"6a0a5723-a022-41f3-9222-c62cac06397e","Type":"ContainerStarted","Data":"06728a7822f65072829b4ddbca0e331708111d1e264540df6b8ad4f8e4b65f68"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.582297 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" event={"ID":"d93236c5-17ab-4cd0-88a3-f4b3479067e5","Type":"ContainerStarted","Data":"69311d84e7163549b83ee0ed47ea75b896b482d3e2b029ab6cc5815195deb913"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.583889 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.587118 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.087102397 +0000 UTC m=+138.380131688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.589665 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" podStartSLOduration=120.589643949 podStartE2EDuration="2m0.589643949s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.535042209 +0000 UTC m=+137.828071502" watchObservedRunningTime="2026-01-21 11:13:15.589643949 +0000 UTC m=+137.882673242" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.589851 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" podStartSLOduration=121.589847152 podStartE2EDuration="2m1.589847152s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.589554391 +0000 UTC m=+137.882583693" watchObservedRunningTime="2026-01-21 11:13:15.589847152 +0000 UTC m=+137.882876443" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.619376 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" event={"ID":"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb","Type":"ContainerStarted","Data":"04ca14e605aa7cca976a17de1d5b8d43161815ffcc069e03b99cab4fbbd6715e"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.628260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" event={"ID":"c873221f-c31d-411a-bbf5-c59e45b473a5","Type":"ContainerStarted","Data":"09c5871e2b1da5897a16dc587de0a6ce894cfabce3a6d3002ba916a166a963d4"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.650597 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" podStartSLOduration=120.650579583 podStartE2EDuration="2m0.650579583s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.641065116 +0000 UTC m=+137.934094408" watchObservedRunningTime="2026-01-21 11:13:15.650579583 +0000 UTC m=+137.943608876" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.688276 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.689314 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.189298797 +0000 UTC m=+138.482328088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.788995 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.789233 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.289207571 +0000 UTC m=+138.582236863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.789659 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.797715 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.297698877 +0000 UTC m=+138.590728170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.894455 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.894987 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.394936465 +0000 UTC m=+138.687965758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.996905 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.997358 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.497337841 +0000 UTC m=+138.790367132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.102395 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.102518 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.602499911 +0000 UTC m=+138.895529203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.102952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.103219 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.603209604 +0000 UTC m=+138.896238896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.163572 4824 csr.go:261] certificate signing request csr-k74p7 is approved, waiting to be issued Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.168747 4824 csr.go:257] certificate signing request csr-k74p7 is issued Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.204716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.204896 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.704868214 +0000 UTC m=+138.997897506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.205005 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.205343 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.705330822 +0000 UTC m=+138.998360115 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.306506 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.306687 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.806664072 +0000 UTC m=+139.099693364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.306950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.307240 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.807230727 +0000 UTC m=+139.100260018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.407856 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.408051 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.908018381 +0000 UTC m=+139.201047673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.408233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.408665 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.908646581 +0000 UTC m=+139.201675872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.465279 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:16 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:16 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:16 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.465349 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.509241 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.509413 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.009388549 +0000 UTC m=+139.302417842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.509647 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.509923 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.009911081 +0000 UTC m=+139.302940373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.611144 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.611340 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.111313721 +0000 UTC m=+139.404343013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.611577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.611917 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.111903788 +0000 UTC m=+139.404933081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.631567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" event={"ID":"26739e4d-344c-4e25-bd0a-019f11456897","Type":"ContainerStarted","Data":"2357fdef6edff16fe248b51b2573129538bac0ebcbe355fd4f0744b4760e074f"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.631624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" event={"ID":"26739e4d-344c-4e25-bd0a-019f11456897","Type":"ContainerStarted","Data":"191c1cbbbb507e726a9fc6a06f49f5bfd133346772551af833babba6bac384fc"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.633052 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" event={"ID":"024ed20d-cf91-4660-998b-3d06d3464607","Type":"ContainerStarted","Data":"8d6da99e5afb06b10adc2e159c9dfb91014cadd7b29865889c55079801523753"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.634187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" event={"ID":"23d7bb25-5d2c-4530-86dd-25a956953a4f","Type":"ContainerStarted","Data":"a18fe41bc9f1a62f11b63c70a3c09ab814ba529ea96b52baf9317658c9a34f69"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.634574 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.635523 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" event={"ID":"b79b805a-7ab1-467f-90a5-ac7944cc564e","Type":"ContainerStarted","Data":"fbd69b8205e4cc5b6a7dc3738ab90446c2c9b31eda354ec6976608d7bb2336a3"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.637923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" event={"ID":"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb","Type":"ContainerStarted","Data":"030ab318b65de947022b4f352c2e1b6768805b8d1dc79b53fdc3a85c031ea89d"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.637969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" event={"ID":"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb","Type":"ContainerStarted","Data":"bd42c74cf5e6dfa746c5d497cdc62153be642693c9ec03faf754c61476ddd331"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.639414 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dlgmw" event={"ID":"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe","Type":"ContainerStarted","Data":"bdf1dfb3e18ae28b5dd5316bda516f1af9158072d9d605c478bfa2943dc87d2a"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.639439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dlgmw" event={"ID":"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe","Type":"ContainerStarted","Data":"2c1919ef7a931735c653fb352fe2a897a5d074c0b408afd192aa4c432f141422"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.639755 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.640728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" event={"ID":"74172695-d17f-4a47-b013-9e025e534916","Type":"ContainerStarted","Data":"98f08cff82a80cfcccb794726fdd3ef6aa3878c820ee87ff8f0ce7a1689c9be4"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.641909 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" event={"ID":"51e27630-d5c1-4eed-8232-3698ab99fc19","Type":"ContainerStarted","Data":"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.643743 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" event={"ID":"3358782c-3435-4c68-9f9c-c1040a6ada32","Type":"ContainerStarted","Data":"33d6f04d21a2f84840ecc147d2a73698c0b7579e93456756c1f45bca7450a4d9"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.644972 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8g4qk" event={"ID":"2279132e-c274-424b-8c8b-80ae5bd7dfe5","Type":"ContainerStarted","Data":"e8a4c509ba9741577881459900939f844d9d1987bece7045fdf2ed555a8af9d8"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.645560 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-8g4qk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.645592 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8g4qk" podUID="2279132e-c274-424b-8c8b-80ae5bd7dfe5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.645748 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.647415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" event={"ID":"8364adda-37b5-4a9d-8abd-3daf12470f20","Type":"ContainerStarted","Data":"e116aa985fb0389e3d80883c44d0f137bc3abbb6bd14ffacfad6761ccd102d60"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.648543 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" event={"ID":"86b5dca2-989f-4b6b-983b-87763ce0d469","Type":"ContainerStarted","Data":"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.648744 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.650383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" event={"ID":"63c09cd8-12ad-40a1-9877-3fc1500acb7a","Type":"ContainerStarted","Data":"2ae4c495da6ee06fbb44a8d0b9889c50c26e09b1bd4eed72119afd916722467d"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.650423 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" event={"ID":"63c09cd8-12ad-40a1-9877-3fc1500acb7a","Type":"ContainerStarted","Data":"fe6496e1452e7435e48ed695e321941578784919e86b59cbda2becfb36ff4c6b"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.650437 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" event={"ID":"63c09cd8-12ad-40a1-9877-3fc1500acb7a","Type":"ContainerStarted","Data":"f05b730bf98627ba801da84010b12c900b72c91b8de345b422941b9fbfff5a3f"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.651607 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" event={"ID":"c873221f-c31d-411a-bbf5-c59e45b473a5","Type":"ContainerStarted","Data":"9876f148fac8d256f8df6efbf5ac5f62b7c3b46dead9a743743d3d5ebb4abb69"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.653599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" event={"ID":"d93236c5-17ab-4cd0-88a3-f4b3479067e5","Type":"ContainerStarted","Data":"766c63224fae3c18fae3adec0594b20bcd6d133f487a645ea28f65cfabad9132"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.653672 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" event={"ID":"d93236c5-17ab-4cd0-88a3-f4b3479067e5","Type":"ContainerStarted","Data":"951567f3adb934892728b522d94d22bc9bc79d7a05baeb1945237817c10327d5"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.655463 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" event={"ID":"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b","Type":"ContainerStarted","Data":"03bc2ab81d22972fd9c8c1c1d652fdedcb038fe3b6031b95a9d65586e643fd12"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.655492 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" event={"ID":"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b","Type":"ContainerStarted","Data":"9b0b5099e430097dfbaa3a46a392984fa7c9bbdf05ca3d28a430160a4137e4d8"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.657032 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b7tsx" event={"ID":"faede743-203b-4bd1-b17a-776a4744bdf1","Type":"ContainerStarted","Data":"934e6b741f99fc948601142b6c4dfe7be70cfbdae26baa682593e0cec9afd503"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.662439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" event={"ID":"7a5d2d02-347e-4799-a28f-f303d51758d0","Type":"ContainerStarted","Data":"eacddd49ef3bc5441741a259a2c6d10efa1a1a68c48c3d11e446b57eb54286ee"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.687904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" event={"ID":"8d0d706a-5fe5-402d-9040-3c763e525554","Type":"ContainerStarted","Data":"8d2b64b2f8dc9c4cadefa7a97077294d8931104ce4823d281735da5165e308e9"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.693053 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gjhp5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.693109 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.712512 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.712943 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.212926766 +0000 UTC m=+139.505956057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.718598 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.732458 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" podStartSLOduration=121.732441063 podStartE2EDuration="2m1.732441063s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.676983576 +0000 UTC m=+138.970012868" watchObservedRunningTime="2026-01-21 11:13:16.732441063 +0000 UTC m=+139.025470355" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.733152 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" podStartSLOduration=122.733146156 podStartE2EDuration="2m2.733146156s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.731538638 +0000 UTC m=+139.024567919" watchObservedRunningTime="2026-01-21 11:13:16.733146156 +0000 UTC m=+139.026175448" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.770180 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-b7tsx" podStartSLOduration=7.770156099 podStartE2EDuration="7.770156099s" podCreationTimestamp="2026-01-21 11:13:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.765145711 +0000 UTC m=+139.058175002" watchObservedRunningTime="2026-01-21 11:13:16.770156099 +0000 UTC m=+139.063185392" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.781184 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" podStartSLOduration=121.781153743 podStartE2EDuration="2m1.781153743s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.780949589 +0000 UTC m=+139.073978881" watchObservedRunningTime="2026-01-21 11:13:16.781153743 +0000 UTC m=+139.074183035" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.813596 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" podStartSLOduration=122.813579416 podStartE2EDuration="2m2.813579416s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.799375462 +0000 UTC m=+139.092404755" watchObservedRunningTime="2026-01-21 11:13:16.813579416 +0000 UTC m=+139.106608708" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.814750 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" podStartSLOduration=122.81474238 podStartE2EDuration="2m2.81474238s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.813641203 +0000 UTC m=+139.106670495" watchObservedRunningTime="2026-01-21 11:13:16.81474238 +0000 UTC m=+139.107771673" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.819043 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.823057 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.3230437 +0000 UTC m=+139.616072993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.831392 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" podStartSLOduration=121.831375126 podStartE2EDuration="2m1.831375126s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.828254035 +0000 UTC m=+139.121283327" watchObservedRunningTime="2026-01-21 11:13:16.831375126 +0000 UTC m=+139.124404418" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.854041 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.854421 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.871021 4824 patch_prober.go:28] interesting pod/apiserver-76f77b778f-htg7q container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]log ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]etcd ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/generic-apiserver-start-informers ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/max-in-flight-filter ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 21 11:13:16 crc kubenswrapper[4824]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/project.openshift.io-projectcache ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-startinformers ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 21 11:13:16 crc kubenswrapper[4824]: livez check failed Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.871058 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" podUID="8364adda-37b5-4a9d-8abd-3daf12470f20" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.891232 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" podStartSLOduration=121.891212789 podStartE2EDuration="2m1.891212789s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.860996194 +0000 UTC m=+139.154025485" watchObservedRunningTime="2026-01-21 11:13:16.891212789 +0000 UTC m=+139.184242071" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.892794 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" podStartSLOduration=121.892783869 podStartE2EDuration="2m1.892783869s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.890288643 +0000 UTC m=+139.183317935" watchObservedRunningTime="2026-01-21 11:13:16.892783869 +0000 UTC m=+139.185813162" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.922561 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.923147 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.923512 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.423497538 +0000 UTC m=+139.716526831 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.952483 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dlgmw" podStartSLOduration=7.952466801 podStartE2EDuration="7.952466801s" podCreationTimestamp="2026-01-21 11:13:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.952439069 +0000 UTC m=+139.245468361" watchObservedRunningTime="2026-01-21 11:13:16.952466801 +0000 UTC m=+139.245496093" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.952573 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" podStartSLOduration=122.952568923 podStartE2EDuration="2m2.952568923s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.927718785 +0000 UTC m=+139.220748067" watchObservedRunningTime="2026-01-21 11:13:16.952568923 +0000 UTC m=+139.245598215" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.969683 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.969736 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.986496 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7q9f2"] Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.987440 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.987601 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.991279 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.995299 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" podStartSLOduration=121.995287777 podStartE2EDuration="2m1.995287777s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.994359353 +0000 UTC m=+139.287388645" watchObservedRunningTime="2026-01-21 11:13:16.995287777 +0000 UTC m=+139.288317069" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.014518 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7q9f2"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.024305 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.024733 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.524722444 +0000 UTC m=+139.817751736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.050458 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" podStartSLOduration=122.050442686 podStartE2EDuration="2m2.050442686s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:17.04928966 +0000 UTC m=+139.342318942" watchObservedRunningTime="2026-01-21 11:13:17.050442686 +0000 UTC m=+139.343471977" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.095847 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" podStartSLOduration=123.095831273 podStartE2EDuration="2m3.095831273s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:17.095016413 +0000 UTC m=+139.388045705" watchObservedRunningTime="2026-01-21 11:13:17.095831273 +0000 UTC m=+139.388860565" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.126312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.126571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.126659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxt8j\" (UniqueName: \"kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.127062 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.627036836 +0000 UTC m=+139.920066128 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.127129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.170741 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-21 11:08:16 +0000 UTC, rotation deadline is 2026-10-14 23:29:23.303812348 +0000 UTC Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.170800 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6396h16m6.133015242s for next certificate rotation Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.192700 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b922x"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.196047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.198490 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.205703 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b922x"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.228371 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.228602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.228527 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxt8j\" (UniqueName: \"kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.228816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.228898 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.229288 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.729275836 +0000 UTC m=+140.022305128 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.229560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.232391 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.250811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxt8j\" (UniqueName: \"kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.320257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.331719 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.332014 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjbt6\" (UniqueName: \"kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.332070 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.332099 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.332214 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.832199001 +0000 UTC m=+140.125228293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.374445 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.375474 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.404027 4824 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.414284 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.433574 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.433611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjbt6\" (UniqueName: \"kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.433671 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.433717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.434192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.434383 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.934364374 +0000 UTC m=+140.227393666 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.434425 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.455745 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjbt6\" (UniqueName: \"kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.467255 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:17 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:17 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:17 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.467315 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.525010 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.534200 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7q9f2"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.534825 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.535121 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd5wx\" (UniqueName: \"kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.535204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.535244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.535354 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.03533927 +0000 UTC m=+140.328368562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: W0121 11:13:17.548205 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b040d3d_f4a0_451a_97de_08bea41c6dce.slice/crio-58f9378707e041408b55a43bb2ea6d9092a260cf0c9b8d81dc167499cf1c4702 WatchSource:0}: Error finding container 58f9378707e041408b55a43bb2ea6d9092a260cf0c9b8d81dc167499cf1c4702: Status 404 returned error can't find the container with id 58f9378707e041408b55a43bb2ea6d9092a260cf0c9b8d81dc167499cf1c4702 Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.577876 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.580454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.581396 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.636568 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd5wx\" (UniqueName: \"kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.636620 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.636656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.636693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.637132 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.637175 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.137145588 +0000 UTC m=+140.430174880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.637348 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.655549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd5wx\" (UniqueName: \"kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.692487 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.695646 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" event={"ID":"26739e4d-344c-4e25-bd0a-019f11456897","Type":"ContainerStarted","Data":"db772caf01df6719f648d34f4586836dd3d53aa61dba88808bffea38b508c47d"} Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.695686 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" event={"ID":"26739e4d-344c-4e25-bd0a-019f11456897","Type":"ContainerStarted","Data":"b59554f789e277dc869c32d73511c732a6fce72010131678c5798f7d7d9489d4"} Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.695697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" event={"ID":"26739e4d-344c-4e25-bd0a-019f11456897","Type":"ContainerStarted","Data":"6585da42ef186bc5fcc0fb1c007fdf2eaf76ee83f09acbf799840ff171450fec"} Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.698195 4824 generic.go:334] "Generic (PLEG): container finished" podID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerID="ca4b551b3e03faea4b176d4748d261efd694766575aea5c759f3fe832098d94c" exitCode=0 Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.698578 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerDied","Data":"ca4b551b3e03faea4b176d4748d261efd694766575aea5c759f3fe832098d94c"} Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.698629 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerStarted","Data":"58f9378707e041408b55a43bb2ea6d9092a260cf0c9b8d81dc167499cf1c4702"} Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.699113 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-8g4qk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.699140 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8g4qk" podUID="2279132e-c274-424b-8c8b-80ae5bd7dfe5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.699796 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.703227 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.703351 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.705236 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b922x"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.705511 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.714508 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" podStartSLOduration=8.712970503 podStartE2EDuration="8.712970503s" podCreationTimestamp="2026-01-21 11:13:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:17.711290378 +0000 UTC m=+140.004319670" watchObservedRunningTime="2026-01-21 11:13:17.712970503 +0000 UTC m=+140.005999795" Jan 21 11:13:17 crc kubenswrapper[4824]: W0121 11:13:17.718674 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f11911f_e365_4151_b536_3988186b8757.slice/crio-8980ea76a1ff2b34ee51f0465162ae2c030d8714d1e864ebb767d3d7a1da8509 WatchSource:0}: Error finding container 8980ea76a1ff2b34ee51f0465162ae2c030d8714d1e864ebb767d3d7a1da8509: Status 404 returned error can't find the container with id 8980ea76a1ff2b34ee51f0465162ae2c030d8714d1e864ebb767d3d7a1da8509 Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.739463 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.739679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtft2\" (UniqueName: \"kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.739713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.739749 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.739932 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.239911587 +0000 UTC m=+140.532940879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.841076 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.841431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtft2\" (UniqueName: \"kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.841505 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.841670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.842180 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.342163623 +0000 UTC m=+140.635192914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.844648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.845185 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.872296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtft2\" (UniqueName: \"kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.902827 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.925800 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.944129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.944532 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.444516638 +0000 UTC m=+140.737545929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.045843 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: E0121 11:13:18.046200 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.546185997 +0000 UTC m=+140.839215290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.127861 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:18 crc kubenswrapper[4824]: W0121 11:13:18.145636 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4205fbde_9c6c_40c8_8c8c_d34a607923e4.slice/crio-f85af8c50c5e5798e08394dc9af278d2e32f05c7d0d8ddcab84b8549c975347a WatchSource:0}: Error finding container f85af8c50c5e5798e08394dc9af278d2e32f05c7d0d8ddcab84b8549c975347a: Status 404 returned error can't find the container with id f85af8c50c5e5798e08394dc9af278d2e32f05c7d0d8ddcab84b8549c975347a Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.148077 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:18 crc kubenswrapper[4824]: E0121 11:13:18.148226 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.648210646 +0000 UTC m=+140.941239937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.148371 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: E0121 11:13:18.148613 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.648603693 +0000 UTC m=+140.941632995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.163242 4824 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-21T11:13:17.404276439Z","Handler":null,"Name":""} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.165351 4824 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.165490 4824 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.249894 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.253772 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.351631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.364371 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.364412 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.382750 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.463617 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:18 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:18 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:18 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.463681 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.658632 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.708240 4824 generic.go:334] "Generic (PLEG): container finished" podID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerID="f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795" exitCode=0 Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.708316 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerDied","Data":"f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.708345 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerStarted","Data":"0fe9a2ce9dc5eb2c88168927f6206b6e2577521e0d1f3b71800ade62c141b6ff"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.725607 4824 generic.go:334] "Generic (PLEG): container finished" podID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerID="12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0" exitCode=0 Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.725689 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerDied","Data":"12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.725719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerStarted","Data":"f85af8c50c5e5798e08394dc9af278d2e32f05c7d0d8ddcab84b8549c975347a"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.734552 4824 generic.go:334] "Generic (PLEG): container finished" podID="3f11911f-e365-4151-b536-3988186b8757" containerID="fe402f640205024f989b1231ac3c21b31c3d7e53bed03337b9aa7afe32b7848e" exitCode=0 Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.734628 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerDied","Data":"fe402f640205024f989b1231ac3c21b31c3d7e53bed03337b9aa7afe32b7848e"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.734656 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerStarted","Data":"8980ea76a1ff2b34ee51f0465162ae2c030d8714d1e864ebb767d3d7a1da8509"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.736529 4824 generic.go:334] "Generic (PLEG): container finished" podID="09bd0b9c-46fb-433f-bd30-014c69c90d9f" containerID="277011d8fed4b1333d53f5a4a084181f52dada1e6d85b31750d9bab9937b3678" exitCode=0 Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.736984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" event={"ID":"09bd0b9c-46fb-433f-bd30-014c69c90d9f","Type":"ContainerDied","Data":"277011d8fed4b1333d53f5a4a084181f52dada1e6d85b31750d9bab9937b3678"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.749077 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.019573 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:13:19 crc kubenswrapper[4824]: W0121 11:13:19.037008 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94e71514_0af6_4254_8815_c933ade6c9da.slice/crio-7d8d535dae1a2ab198548905f9a9571895c25a2b895cfb327e93cc9833852054 WatchSource:0}: Error finding container 7d8d535dae1a2ab198548905f9a9571895c25a2b895cfb327e93cc9833852054: Status 404 returned error can't find the container with id 7d8d535dae1a2ab198548905f9a9571895c25a2b895cfb327e93cc9833852054 Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.168906 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4w5m6"] Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.171386 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.175064 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.176626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4w5m6"] Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.365365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.365441 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.365499 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lchhw\" (UniqueName: \"kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.463544 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:19 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:19 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:19 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.463601 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.474922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.475302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lchhw\" (UniqueName: \"kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.475970 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.475887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.476204 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.515278 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lchhw\" (UniqueName: \"kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.573161 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.576621 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.584715 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.678738 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.679050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.679241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tskqq\" (UniqueName: \"kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.752400 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" event={"ID":"94e71514-0af6-4254-8815-c933ade6c9da","Type":"ContainerStarted","Data":"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf"} Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.752444 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" event={"ID":"94e71514-0af6-4254-8815-c933ade6c9da","Type":"ContainerStarted","Data":"7d8d535dae1a2ab198548905f9a9571895c25a2b895cfb327e93cc9833852054"} Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.752620 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.786208 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" podStartSLOduration=124.786187885 podStartE2EDuration="2m4.786187885s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:19.781783293 +0000 UTC m=+142.074812585" watchObservedRunningTime="2026-01-21 11:13:19.786187885 +0000 UTC m=+142.079217177" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.788937 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.792071 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.792189 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.792404 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tskqq\" (UniqueName: \"kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.792707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.793335 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.809251 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tskqq\" (UniqueName: \"kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.894580 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.081361 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.082258 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4w5m6"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.090120 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:20 crc kubenswrapper[4824]: W0121 11:13:20.093176 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bc8c652_fb4f_48b7_86ce_c45522dfd476.slice/crio-1717bf3078cfbe6fee3800d069a35c471757cb03a1eddb1d8ff1e20167f89c9e WatchSource:0}: Error finding container 1717bf3078cfbe6fee3800d069a35c471757cb03a1eddb1d8ff1e20167f89c9e: Status 404 returned error can't find the container with id 1717bf3078cfbe6fee3800d069a35c471757cb03a1eddb1d8ff1e20167f89c9e Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.142740 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 21 11:13:20 crc kubenswrapper[4824]: E0121 11:13:20.143051 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09bd0b9c-46fb-433f-bd30-014c69c90d9f" containerName="collect-profiles" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.143065 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="09bd0b9c-46fb-433f-bd30-014c69c90d9f" containerName="collect-profiles" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.143185 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="09bd0b9c-46fb-433f-bd30-014c69c90d9f" containerName="collect-profiles" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.143570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.145269 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.146301 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.146400 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.168665 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hqs52"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.169681 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.171704 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.177276 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hqs52"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.199361 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume\") pod \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.199393 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume\") pod \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.200260 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqqww\" (UniqueName: \"kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww\") pod \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.200245 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume" (OuterVolumeSpecName: "config-volume") pod "09bd0b9c-46fb-433f-bd30-014c69c90d9f" (UID: "09bd0b9c-46fb-433f-bd30-014c69c90d9f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.205999 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "09bd0b9c-46fb-433f-bd30-014c69c90d9f" (UID: "09bd0b9c-46fb-433f-bd30-014c69c90d9f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.206131 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww" (OuterVolumeSpecName: "kube-api-access-nqqww") pod "09bd0b9c-46fb-433f-bd30-014c69c90d9f" (UID: "09bd0b9c-46fb-433f-bd30-014c69c90d9f"). InnerVolumeSpecName "kube-api-access-nqqww". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.302182 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.302512 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.302701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.302805 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8656b\" (UniqueName: \"kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.302842 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.303011 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqqww\" (UniqueName: \"kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.303040 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.303052 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.374020 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:20 crc kubenswrapper[4824]: W0121 11:13:20.393748 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff7a0886_d523_486a_9bf0_bae3456b85c2.slice/crio-1cc0a8fad4ae405d30baab5682d5e3fd0a25f77c99ddd8f27c737f65b8432c15 WatchSource:0}: Error finding container 1cc0a8fad4ae405d30baab5682d5e3fd0a25f77c99ddd8f27c737f65b8432c15: Status 404 returned error can't find the container with id 1cc0a8fad4ae405d30baab5682d5e3fd0a25f77c99ddd8f27c737f65b8432c15 Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.403877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8656b\" (UniqueName: \"kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.403942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.403983 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.404003 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.404070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.404447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.405327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.406635 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.417855 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8656b\" (UniqueName: \"kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.420539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.462160 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.463673 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:20 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:20 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:20 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.463724 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.495435 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.572554 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.577786 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.584608 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.611595 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.611678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcjz2\" (UniqueName: \"kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.613849 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.671716 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.715519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcjz2\" (UniqueName: \"kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.715598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.715661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.716499 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.716616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.716994 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hqs52"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.731847 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcjz2\" (UniqueName: \"kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.760842 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.760846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" event={"ID":"09bd0b9c-46fb-433f-bd30-014c69c90d9f","Type":"ContainerDied","Data":"68b9bb179c8bbe2beb17e913adc7df65d063a4b5a6cdcd18f32da44f14fdb4ae"} Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.760879 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68b9bb179c8bbe2beb17e913adc7df65d063a4b5a6cdcd18f32da44f14fdb4ae" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.767805 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerID="a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3" exitCode=0 Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.767908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerDied","Data":"a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3"} Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.767973 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerStarted","Data":"1cc0a8fad4ae405d30baab5682d5e3fd0a25f77c99ddd8f27c737f65b8432c15"} Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.775408 4824 generic.go:334] "Generic (PLEG): container finished" podID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerID="d5b0c30b7ab307dc38d2c9ff03ad628bf5226a1bb36f64f546ca1dedef2c81fc" exitCode=0 Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.775450 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerDied","Data":"d5b0c30b7ab307dc38d2c9ff03ad628bf5226a1bb36f64f546ca1dedef2c81fc"} Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.775475 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerStarted","Data":"1717bf3078cfbe6fee3800d069a35c471757cb03a1eddb1d8ff1e20167f89c9e"} Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.897484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.463672 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:21 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:21 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:21 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.463724 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.859685 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.863461 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.963047 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.963093 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.964515 4824 patch_prober.go:28] interesting pod/console-f9d7485db-nwlwd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.964559 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-nwlwd" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.411828 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.412743 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.414933 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.416025 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.421228 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.460765 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.462768 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:22 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:22 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:22 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.462819 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.549902 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.550055 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.651386 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.651465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.651495 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.670306 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.683671 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.729169 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.958219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.958266 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.959293 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.973123 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.060322 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.060396 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.063033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.066792 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.263089 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.358408 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.362237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.462898 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:23 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:23 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:23 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.462968 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:23 crc kubenswrapper[4824]: W0121 11:13:23.915586 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3b56f2d_96c0_4597_9235_147fc4802d19.slice/crio-331b281d41eb6e081cb5f5178cf5cdf5ccdc6fe5086d467942c51189e3b27cd1 WatchSource:0}: Error finding container 331b281d41eb6e081cb5f5178cf5cdf5ccdc6fe5086d467942c51189e3b27cd1: Status 404 returned error can't find the container with id 331b281d41eb6e081cb5f5178cf5cdf5ccdc6fe5086d467942c51189e3b27cd1 Jan 21 11:13:23 crc kubenswrapper[4824]: W0121 11:13:23.917622 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5b5068e5_8cc3_4ed1_811c_835a08eefaa0.slice/crio-7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440 WatchSource:0}: Error finding container 7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440: Status 404 returned error can't find the container with id 7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440 Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.468271 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:24 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:24 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:24 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.468321 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:24 crc kubenswrapper[4824]: W0121 11:13:24.777644 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-2ab8a4a5ecc0fa65f52a10055dc79fca26c29a2c4809d5a9e1a0b673c37f378d WatchSource:0}: Error finding container 2ab8a4a5ecc0fa65f52a10055dc79fca26c29a2c4809d5a9e1a0b673c37f378d: Status 404 returned error can't find the container with id 2ab8a4a5ecc0fa65f52a10055dc79fca26c29a2c4809d5a9e1a0b673c37f378d Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.800192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2ab8a4a5ecc0fa65f52a10055dc79fca26c29a2c4809d5a9e1a0b673c37f378d"} Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.808263 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5b5068e5-8cc3-4ed1-811c-835a08eefaa0","Type":"ContainerStarted","Data":"7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440"} Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.810907 4824 generic.go:334] "Generic (PLEG): container finished" podID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerID="550d4757a2a4f775989081e05280aba8d60bf329b7f947d0664b2247319cf4a8" exitCode=0 Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.811000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerDied","Data":"550d4757a2a4f775989081e05280aba8d60bf329b7f947d0664b2247319cf4a8"} Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.811045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerStarted","Data":"331b281d41eb6e081cb5f5178cf5cdf5ccdc6fe5086d467942c51189e3b27cd1"} Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.818978 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"00d0630de43ad5fd4cd1bf06c27ca7ff6e3a12d85248fbd78a804e3acd633ef7"} Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.820171 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.915620 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:24 crc kubenswrapper[4824]: W0121 11:13:24.922239 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30363b0b_09e8_4fb6_88e4_4e941d833ec8.slice/crio-e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace WatchSource:0}: Error finding container e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace: Status 404 returned error can't find the container with id e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.099044 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.463356 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:25 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:25 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:25 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.463575 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.829236 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a640fdf54e10f37059bf9750e724548215efaf4e93fd9602296529be005be06f"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.839051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"02e04011427f72355c3b733fee5ade53165d5785d1fa946743d9fba748deca4e"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.849261 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b5068e5-8cc3-4ed1-811c-835a08eefaa0" containerID="ef6e69b85db227503d6edb2ea4085936ac71a7ca73f8fe10106c4a95a99e875d" exitCode=0 Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.849472 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5b5068e5-8cc3-4ed1-811c-835a08eefaa0","Type":"ContainerDied","Data":"ef6e69b85db227503d6edb2ea4085936ac71a7ca73f8fe10106c4a95a99e875d"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.859645 4824 generic.go:334] "Generic (PLEG): container finished" podID="5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" containerID="7a44ff0047a3d569ff3920eabb699033dff3d6825b4197844557de3ddee34532" exitCode=0 Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.860069 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec","Type":"ContainerDied","Data":"7a44ff0047a3d569ff3920eabb699033dff3d6825b4197844557de3ddee34532"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.860113 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec","Type":"ContainerStarted","Data":"e40b90f30056636be3e795d67f23ef53655a16ac6e0047bd3710d44978191f9c"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.867385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9e75f6e09ca96c00083080187868f0959b24c9bec3aa61af4b5463f9160f2778"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.867411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"67a4de3580d12468ade2e60b1a645d0031099e3f491957f9a01eb4b06f057b99"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.867587 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.868664 4824 generic.go:334] "Generic (PLEG): container finished" podID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerID="75d2b0ba70b6aa340dd2f71ac90a2f4234cfa6ec869faff5f999306922a53203" exitCode=0 Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.868692 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerDied","Data":"75d2b0ba70b6aa340dd2f71ac90a2f4234cfa6ec869faff5f999306922a53203"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.869215 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerStarted","Data":"e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace"} Jan 21 11:13:26 crc kubenswrapper[4824]: I0121 11:13:26.462679 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:26 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:26 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:26 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:26 crc kubenswrapper[4824]: I0121 11:13:26.462928 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:27 crc kubenswrapper[4824]: I0121 11:13:27.462914 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:27 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:27 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:27 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:27 crc kubenswrapper[4824]: I0121 11:13:27.462990 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:28 crc kubenswrapper[4824]: I0121 11:13:28.462744 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:28 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:28 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:28 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:28 crc kubenswrapper[4824]: I0121 11:13:28.462981 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:28 crc kubenswrapper[4824]: I0121 11:13:28.750716 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.462732 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:29 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:29 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:29 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.462788 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.790141 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.854802 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir\") pod \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.854849 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access\") pod \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.854989 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5b5068e5-8cc3-4ed1-811c-835a08eefaa0" (UID: "5b5068e5-8cc3-4ed1-811c-835a08eefaa0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.855162 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.859761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5b5068e5-8cc3-4ed1-811c-835a08eefaa0" (UID: "5b5068e5-8cc3-4ed1-811c-835a08eefaa0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.913159 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5b5068e5-8cc3-4ed1-811c-835a08eefaa0","Type":"ContainerDied","Data":"7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440"} Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.913193 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.913240 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.956708 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:30 crc kubenswrapper[4824]: I0121 11:13:30.462435 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:30 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:30 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:30 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:30 crc kubenswrapper[4824]: I0121 11:13:30.462492 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.050031 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.068561 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir\") pod \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.068637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access\") pod \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.069240 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" (UID: "5a6cefc5-7c75-40ca-b14e-d0f7615b7cec"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.072828 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" (UID: "5a6cefc5-7c75-40ca-b14e-d0f7615b7cec"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.170124 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.170646 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.462833 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.465386 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.923473 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.928869 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec","Type":"ContainerDied","Data":"e40b90f30056636be3e795d67f23ef53655a16ac6e0047bd3710d44978191f9c"} Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.928930 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e40b90f30056636be3e795d67f23ef53655a16ac6e0047bd3710d44978191f9c" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.963934 4824 patch_prober.go:28] interesting pod/console-f9d7485db-nwlwd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.963986 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-nwlwd" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 21 11:13:35 crc kubenswrapper[4824]: I0121 11:13:35.943010 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerStarted","Data":"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06"} Jan 21 11:13:35 crc kubenswrapper[4824]: I0121 11:13:35.944982 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerStarted","Data":"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.239355 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.244334 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.458417 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.793574 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9rmnw"] Jan 21 11:13:36 crc kubenswrapper[4824]: W0121 11:13:36.797271 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f9bd12b_22ee_44ed_a8b9_c100d043f691.slice/crio-7ca13b06e8160aab5b17fb64bf1b61c7509021a58c16e7661855aeaab33e7d91 WatchSource:0}: Error finding container 7ca13b06e8160aab5b17fb64bf1b61c7509021a58c16e7661855aeaab33e7d91: Status 404 returned error can't find the container with id 7ca13b06e8160aab5b17fb64bf1b61c7509021a58c16e7661855aeaab33e7d91 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.950641 4824 generic.go:334] "Generic (PLEG): container finished" podID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerID="0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.950736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerDied","Data":"0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.952462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" event={"ID":"8f9bd12b-22ee-44ed-a8b9-c100d043f691","Type":"ContainerStarted","Data":"90841fa6243d7a074f0ac613f01a7d1e84fde5829aa9cb007204f2ee1f35fe14"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.952498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" event={"ID":"8f9bd12b-22ee-44ed-a8b9-c100d043f691","Type":"ContainerStarted","Data":"7ca13b06e8160aab5b17fb64bf1b61c7509021a58c16e7661855aeaab33e7d91"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.954777 4824 generic.go:334] "Generic (PLEG): container finished" podID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerID="122ae0144dee544f8ce79c06f524513fd679f1334f79370a0f9e8b920f8d7ee4" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.954842 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerDied","Data":"122ae0144dee544f8ce79c06f524513fd679f1334f79370a0f9e8b920f8d7ee4"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.963833 4824 generic.go:334] "Generic (PLEG): container finished" podID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerID="e964ec9dc2f3dad14f23b05ef15becda9a72c91eb5ec8bd728789277a97681be" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.963917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerDied","Data":"e964ec9dc2f3dad14f23b05ef15becda9a72c91eb5ec8bd728789277a97681be"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.968455 4824 generic.go:334] "Generic (PLEG): container finished" podID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerID="9f778c1f1c308a32a3e2af072eab8265d05e159ea1a783308d30e563350f290a" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.968532 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerDied","Data":"9f778c1f1c308a32a3e2af072eab8265d05e159ea1a783308d30e563350f290a"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.977001 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerID="294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.977542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerDied","Data":"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.982678 4824 generic.go:334] "Generic (PLEG): container finished" podID="3f11911f-e365-4151-b536-3988186b8757" containerID="bc5314fd2d078e4f04072ac4dda373754c451434b3d4dce44df1b40cbeecd1ba" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.982733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerDied","Data":"bc5314fd2d078e4f04072ac4dda373754c451434b3d4dce44df1b40cbeecd1ba"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.985525 4824 generic.go:334] "Generic (PLEG): container finished" podID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerID="5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.985572 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerDied","Data":"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.988417 4824 generic.go:334] "Generic (PLEG): container finished" podID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerID="2f3db218d18e31256264bc1967baec36484f60d00de32760d6c6503fe4e7b9fc" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.988449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerDied","Data":"2f3db218d18e31256264bc1967baec36484f60d00de32760d6c6503fe4e7b9fc"} Jan 21 11:13:37 crc kubenswrapper[4824]: I0121 11:13:37.996799 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerStarted","Data":"fa2ad6f2a50b9350a988947c908be8adb082c0d1e40507b63b02b300f1699aae"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.000005 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerStarted","Data":"812b638a67ec88d3b10766fb23536e55485cc458e7ca4b815bcac4f9331b7868"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.007879 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerStarted","Data":"f4386ff52380b3c795405ecbb9387d4d9b5fca2cffa267eb0ddb32961b0a7f10"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.009917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerStarted","Data":"728ea008f822360aaa1493d68beaf88c96b2cdf76a7eeb252c13acb6bf91c9d5"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.018181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerStarted","Data":"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.020804 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerStarted","Data":"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.022451 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" event={"ID":"8f9bd12b-22ee-44ed-a8b9-c100d043f691","Type":"ContainerStarted","Data":"8232a034b266977b397d20f7cdf89da9cd11f33af94ea1374cea91d5d9736037"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.024407 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerStarted","Data":"b1068da5b4766f19c34535e8e58a22325fb8a2d31ce06ee547c973539c887a47"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.033531 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b922x" podStartSLOduration=2.191925607 podStartE2EDuration="21.033521717s" podCreationTimestamp="2026-01-21 11:13:17 +0000 UTC" firstStartedPulling="2026-01-21 11:13:18.73587321 +0000 UTC m=+141.028902502" lastFinishedPulling="2026-01-21 11:13:37.57746932 +0000 UTC m=+159.870498612" observedRunningTime="2026-01-21 11:13:38.031985582 +0000 UTC m=+160.325014874" watchObservedRunningTime="2026-01-21 11:13:38.033521717 +0000 UTC m=+160.326551009" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.034130 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4w5m6" podStartSLOduration=5.39873712 podStartE2EDuration="19.03412513s" podCreationTimestamp="2026-01-21 11:13:19 +0000 UTC" firstStartedPulling="2026-01-21 11:13:23.904887702 +0000 UTC m=+146.197916994" lastFinishedPulling="2026-01-21 11:13:37.540275712 +0000 UTC m=+159.833305004" observedRunningTime="2026-01-21 11:13:38.016888951 +0000 UTC m=+160.309918243" watchObservedRunningTime="2026-01-21 11:13:38.03412513 +0000 UTC m=+160.327154422" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.048906 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2qhk6" podStartSLOduration=6.245616317 podStartE2EDuration="18.048892321s" podCreationTimestamp="2026-01-21 11:13:20 +0000 UTC" firstStartedPulling="2026-01-21 11:13:25.870187028 +0000 UTC m=+148.163216320" lastFinishedPulling="2026-01-21 11:13:37.673463032 +0000 UTC m=+159.966492324" observedRunningTime="2026-01-21 11:13:38.046673826 +0000 UTC m=+160.339703118" watchObservedRunningTime="2026-01-21 11:13:38.048892321 +0000 UTC m=+160.341921613" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.064211 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7q9f2" podStartSLOduration=2.232781756 podStartE2EDuration="22.064192926s" podCreationTimestamp="2026-01-21 11:13:16 +0000 UTC" firstStartedPulling="2026-01-21 11:13:17.702937212 +0000 UTC m=+139.995966504" lastFinishedPulling="2026-01-21 11:13:37.534348381 +0000 UTC m=+159.827377674" observedRunningTime="2026-01-21 11:13:38.062946625 +0000 UTC m=+160.355975918" watchObservedRunningTime="2026-01-21 11:13:38.064192926 +0000 UTC m=+160.357222218" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.102187 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pckcl" podStartSLOduration=2.284079136 podStartE2EDuration="21.10217267s" podCreationTimestamp="2026-01-21 11:13:17 +0000 UTC" firstStartedPulling="2026-01-21 11:13:18.710219794 +0000 UTC m=+141.003249086" lastFinishedPulling="2026-01-21 11:13:37.528313328 +0000 UTC m=+159.821342620" observedRunningTime="2026-01-21 11:13:38.101510446 +0000 UTC m=+160.394539738" watchObservedRunningTime="2026-01-21 11:13:38.10217267 +0000 UTC m=+160.395201962" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.102564 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-9rmnw" podStartSLOduration=144.102559767 podStartE2EDuration="2m24.102559767s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:38.085539363 +0000 UTC m=+160.378568656" watchObservedRunningTime="2026-01-21 11:13:38.102559767 +0000 UTC m=+160.395589059" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.118160 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hqs52" podStartSLOduration=5.235102315 podStartE2EDuration="18.118145125s" podCreationTimestamp="2026-01-21 11:13:20 +0000 UTC" firstStartedPulling="2026-01-21 11:13:24.819519941 +0000 UTC m=+147.112549233" lastFinishedPulling="2026-01-21 11:13:37.702562751 +0000 UTC m=+159.995592043" observedRunningTime="2026-01-21 11:13:38.114780748 +0000 UTC m=+160.407810039" watchObservedRunningTime="2026-01-21 11:13:38.118145125 +0000 UTC m=+160.411174418" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.137792 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s2hsr" podStartSLOduration=2.254256471 podStartE2EDuration="21.137741697s" podCreationTimestamp="2026-01-21 11:13:17 +0000 UTC" firstStartedPulling="2026-01-21 11:13:18.72741732 +0000 UTC m=+141.020446612" lastFinishedPulling="2026-01-21 11:13:37.610902546 +0000 UTC m=+159.903931838" observedRunningTime="2026-01-21 11:13:38.13755691 +0000 UTC m=+160.430586202" watchObservedRunningTime="2026-01-21 11:13:38.137741697 +0000 UTC m=+160.430770989" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.664346 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.046702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerStarted","Data":"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd"} Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.062824 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8dj94" podStartSLOduration=6.064740655 podStartE2EDuration="20.062811307s" podCreationTimestamp="2026-01-21 11:13:19 +0000 UTC" firstStartedPulling="2026-01-21 11:13:23.90538185 +0000 UTC m=+146.198411142" lastFinishedPulling="2026-01-21 11:13:37.903452502 +0000 UTC m=+160.196481794" observedRunningTime="2026-01-21 11:13:39.061787946 +0000 UTC m=+161.354817238" watchObservedRunningTime="2026-01-21 11:13:39.062811307 +0000 UTC m=+161.355840599" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.789575 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.789622 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.859098 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.895181 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.895225 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.922491 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:40 crc kubenswrapper[4824]: I0121 11:13:40.495766 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:40 crc kubenswrapper[4824]: I0121 11:13:40.496233 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:40 crc kubenswrapper[4824]: I0121 11:13:40.898144 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:40 crc kubenswrapper[4824]: I0121 11:13:40.898188 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:41 crc kubenswrapper[4824]: I0121 11:13:41.522458 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hqs52" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="registry-server" probeResult="failure" output=< Jan 21 11:13:41 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Jan 21 11:13:41 crc kubenswrapper[4824]: > Jan 21 11:13:41 crc kubenswrapper[4824]: I0121 11:13:41.923566 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2qhk6" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="registry-server" probeResult="failure" output=< Jan 21 11:13:41 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Jan 21 11:13:41 crc kubenswrapper[4824]: > Jan 21 11:13:41 crc kubenswrapper[4824]: I0121 11:13:41.968742 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:41 crc kubenswrapper[4824]: I0121 11:13:41.972070 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:46 crc kubenswrapper[4824]: I0121 11:13:46.065537 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:13:46 crc kubenswrapper[4824]: I0121 11:13:46.065794 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.321543 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.322082 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.352502 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.525514 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.525745 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.553045 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.693565 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.693792 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.720050 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.904235 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.904427 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.931122 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:48 crc kubenswrapper[4824]: I0121 11:13:48.107713 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:48 crc kubenswrapper[4824]: I0121 11:13:48.108143 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:48 crc kubenswrapper[4824]: I0121 11:13:48.109053 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:48 crc kubenswrapper[4824]: I0121 11:13:48.109185 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:49 crc kubenswrapper[4824]: I0121 11:13:49.375113 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:49 crc kubenswrapper[4824]: I0121 11:13:49.816095 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:49 crc kubenswrapper[4824]: I0121 11:13:49.921485 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:49 crc kubenswrapper[4824]: I0121 11:13:49.978224 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.092539 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s2hsr" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="registry-server" containerID="cri-o://ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395" gracePeriod=2 Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.450833 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.523193 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.528033 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content\") pod \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.528073 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities\") pod \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.528093 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtft2\" (UniqueName: \"kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2\") pod \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.529357 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities" (OuterVolumeSpecName: "utilities") pod "4205fbde-9c6c-40c8-8c8c-d34a607923e4" (UID: "4205fbde-9c6c-40c8-8c8c-d34a607923e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.532787 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2" (OuterVolumeSpecName: "kube-api-access-mtft2") pod "4205fbde-9c6c-40c8-8c8c-d34a607923e4" (UID: "4205fbde-9c6c-40c8-8c8c-d34a607923e4"). InnerVolumeSpecName "kube-api-access-mtft2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.550779 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.575149 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4205fbde-9c6c-40c8-8c8c-d34a607923e4" (UID: "4205fbde-9c6c-40c8-8c8c-d34a607923e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.629231 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.629262 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.629274 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtft2\" (UniqueName: \"kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.924674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.954488 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.098037 4824 generic.go:334] "Generic (PLEG): container finished" podID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerID="ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395" exitCode=0 Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.098618 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.105876 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerDied","Data":"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395"} Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.105921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerDied","Data":"f85af8c50c5e5798e08394dc9af278d2e32f05c7d0d8ddcab84b8549c975347a"} Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.105986 4824 scope.go:117] "RemoveContainer" containerID="ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.106091 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pckcl" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="registry-server" containerID="cri-o://89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7" gracePeriod=2 Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.119214 4824 scope.go:117] "RemoveContainer" containerID="5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.125805 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.127800 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.159842 4824 scope.go:117] "RemoveContainer" containerID="12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.208760 4824 scope.go:117] "RemoveContainer" containerID="ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395" Jan 21 11:13:51 crc kubenswrapper[4824]: E0121 11:13:51.209139 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395\": container with ID starting with ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395 not found: ID does not exist" containerID="ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.209186 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395"} err="failed to get container status \"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395\": rpc error: code = NotFound desc = could not find container \"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395\": container with ID starting with ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395 not found: ID does not exist" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.209229 4824 scope.go:117] "RemoveContainer" containerID="5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543" Jan 21 11:13:51 crc kubenswrapper[4824]: E0121 11:13:51.209534 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543\": container with ID starting with 5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543 not found: ID does not exist" containerID="5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.209560 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543"} err="failed to get container status \"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543\": rpc error: code = NotFound desc = could not find container \"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543\": container with ID starting with 5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543 not found: ID does not exist" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.209587 4824 scope.go:117] "RemoveContainer" containerID="12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0" Jan 21 11:13:51 crc kubenswrapper[4824]: E0121 11:13:51.209794 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0\": container with ID starting with 12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0 not found: ID does not exist" containerID="12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.209817 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0"} err="failed to get container status \"12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0\": rpc error: code = NotFound desc = could not find container \"12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0\": container with ID starting with 12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0 not found: ID does not exist" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.457751 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.543162 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities\") pod \"0939eda4-fff7-447b-ad52-ddb427bf1272\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.543404 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd5wx\" (UniqueName: \"kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx\") pod \"0939eda4-fff7-447b-ad52-ddb427bf1272\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.543552 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content\") pod \"0939eda4-fff7-447b-ad52-ddb427bf1272\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.543839 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities" (OuterVolumeSpecName: "utilities") pod "0939eda4-fff7-447b-ad52-ddb427bf1272" (UID: "0939eda4-fff7-447b-ad52-ddb427bf1272"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.546246 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx" (OuterVolumeSpecName: "kube-api-access-gd5wx") pod "0939eda4-fff7-447b-ad52-ddb427bf1272" (UID: "0939eda4-fff7-447b-ad52-ddb427bf1272"). InnerVolumeSpecName "kube-api-access-gd5wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.580998 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0939eda4-fff7-447b-ad52-ddb427bf1272" (UID: "0939eda4-fff7-447b-ad52-ddb427bf1272"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.649312 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.649344 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd5wx\" (UniqueName: \"kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.649355 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.055095 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" path="/var/lib/kubelet/pods/4205fbde-9c6c-40c8-8c8c-d34a607923e4/volumes" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.103652 4824 generic.go:334] "Generic (PLEG): container finished" podID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerID="89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7" exitCode=0 Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.103685 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerDied","Data":"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7"} Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.103708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerDied","Data":"0fe9a2ce9dc5eb2c88168927f6206b6e2577521e0d1f3b71800ade62c141b6ff"} Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.103716 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.103724 4824 scope.go:117] "RemoveContainer" containerID="89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.117384 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.117412 4824 scope.go:117] "RemoveContainer" containerID="0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.118198 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.128032 4824 scope.go:117] "RemoveContainer" containerID="f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.136661 4824 scope.go:117] "RemoveContainer" containerID="89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7" Jan 21 11:13:52 crc kubenswrapper[4824]: E0121 11:13:52.136981 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7\": container with ID starting with 89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7 not found: ID does not exist" containerID="89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.137072 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7"} err="failed to get container status \"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7\": rpc error: code = NotFound desc = could not find container \"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7\": container with ID starting with 89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7 not found: ID does not exist" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.137142 4824 scope.go:117] "RemoveContainer" containerID="0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea" Jan 21 11:13:52 crc kubenswrapper[4824]: E0121 11:13:52.137422 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea\": container with ID starting with 0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea not found: ID does not exist" containerID="0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.137501 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea"} err="failed to get container status \"0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea\": rpc error: code = NotFound desc = could not find container \"0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea\": container with ID starting with 0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea not found: ID does not exist" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.137561 4824 scope.go:117] "RemoveContainer" containerID="f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795" Jan 21 11:13:52 crc kubenswrapper[4824]: E0121 11:13:52.137840 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795\": container with ID starting with f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795 not found: ID does not exist" containerID="f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.137868 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795"} err="failed to get container status \"f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795\": rpc error: code = NotFound desc = could not find container \"f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795\": container with ID starting with f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795 not found: ID does not exist" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.376395 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.376790 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8dj94" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="registry-server" containerID="cri-o://963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd" gracePeriod=2 Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.421331 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.728693 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.862555 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tskqq\" (UniqueName: \"kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq\") pod \"ff7a0886-d523-486a-9bf0-bae3456b85c2\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.862679 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities\") pod \"ff7a0886-d523-486a-9bf0-bae3456b85c2\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.862717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content\") pod \"ff7a0886-d523-486a-9bf0-bae3456b85c2\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.868347 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq" (OuterVolumeSpecName: "kube-api-access-tskqq") pod "ff7a0886-d523-486a-9bf0-bae3456b85c2" (UID: "ff7a0886-d523-486a-9bf0-bae3456b85c2"). InnerVolumeSpecName "kube-api-access-tskqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.877979 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities" (OuterVolumeSpecName: "utilities") pod "ff7a0886-d523-486a-9bf0-bae3456b85c2" (UID: "ff7a0886-d523-486a-9bf0-bae3456b85c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.879550 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff7a0886-d523-486a-9bf0-bae3456b85c2" (UID: "ff7a0886-d523-486a-9bf0-bae3456b85c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.963638 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.963661 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.963673 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tskqq\" (UniqueName: \"kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.108839 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerID="963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd" exitCode=0 Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.108898 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.108905 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerDied","Data":"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd"} Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.108984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerDied","Data":"1cc0a8fad4ae405d30baab5682d5e3fd0a25f77c99ddd8f27c737f65b8432c15"} Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.109012 4824 scope.go:117] "RemoveContainer" containerID="963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.120551 4824 scope.go:117] "RemoveContainer" containerID="294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.130005 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.132231 4824 scope.go:117] "RemoveContainer" containerID="a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.132316 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.159155 4824 scope.go:117] "RemoveContainer" containerID="963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd" Jan 21 11:13:53 crc kubenswrapper[4824]: E0121 11:13:53.159449 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd\": container with ID starting with 963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd not found: ID does not exist" containerID="963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.159478 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd"} err="failed to get container status \"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd\": rpc error: code = NotFound desc = could not find container \"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd\": container with ID starting with 963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd not found: ID does not exist" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.159502 4824 scope.go:117] "RemoveContainer" containerID="294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06" Jan 21 11:13:53 crc kubenswrapper[4824]: E0121 11:13:53.159758 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06\": container with ID starting with 294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06 not found: ID does not exist" containerID="294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.159781 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06"} err="failed to get container status \"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06\": rpc error: code = NotFound desc = could not find container \"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06\": container with ID starting with 294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06 not found: ID does not exist" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.159793 4824 scope.go:117] "RemoveContainer" containerID="a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3" Jan 21 11:13:53 crc kubenswrapper[4824]: E0121 11:13:53.160297 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3\": container with ID starting with a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3 not found: ID does not exist" containerID="a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.160315 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3"} err="failed to get container status \"a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3\": rpc error: code = NotFound desc = could not find container \"a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3\": container with ID starting with a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3 not found: ID does not exist" Jan 21 11:13:54 crc kubenswrapper[4824]: I0121 11:13:54.054048 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" path="/var/lib/kubelet/pods/0939eda4-fff7-447b-ad52-ddb427bf1272/volumes" Jan 21 11:13:54 crc kubenswrapper[4824]: I0121 11:13:54.054570 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" path="/var/lib/kubelet/pods/ff7a0886-d523-486a-9bf0-bae3456b85c2/volumes" Jan 21 11:13:54 crc kubenswrapper[4824]: I0121 11:13:54.774562 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:54 crc kubenswrapper[4824]: I0121 11:13:54.774788 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2qhk6" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="registry-server" containerID="cri-o://812b638a67ec88d3b10766fb23536e55485cc458e7ca4b815bcac4f9331b7868" gracePeriod=2 Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.123372 4824 generic.go:334] "Generic (PLEG): container finished" podID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerID="812b638a67ec88d3b10766fb23536e55485cc458e7ca4b815bcac4f9331b7868" exitCode=0 Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.123411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerDied","Data":"812b638a67ec88d3b10766fb23536e55485cc458e7ca4b815bcac4f9331b7868"} Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.123432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerDied","Data":"e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace"} Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.123442 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.142142 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.193361 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcjz2\" (UniqueName: \"kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2\") pod \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.193453 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content\") pod \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.193482 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities\") pod \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.194280 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities" (OuterVolumeSpecName: "utilities") pod "30363b0b-09e8-4fb6-88e4-4e941d833ec8" (UID: "30363b0b-09e8-4fb6-88e4-4e941d833ec8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.198068 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2" (OuterVolumeSpecName: "kube-api-access-xcjz2") pod "30363b0b-09e8-4fb6-88e4-4e941d833ec8" (UID: "30363b0b-09e8-4fb6-88e4-4e941d833ec8"). InnerVolumeSpecName "kube-api-access-xcjz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.288457 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30363b0b-09e8-4fb6-88e4-4e941d833ec8" (UID: "30363b0b-09e8-4fb6-88e4-4e941d833ec8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.294256 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.294276 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.294286 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcjz2\" (UniqueName: \"kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:56 crc kubenswrapper[4824]: I0121 11:13:56.127175 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:56 crc kubenswrapper[4824]: I0121 11:13:56.139286 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:56 crc kubenswrapper[4824]: I0121 11:13:56.141565 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.053290 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" path="/var/lib/kubelet/pods/30363b0b-09e8-4fb6-88e4-4e941d833ec8/volumes" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.603991 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604319 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604336 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604348 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b5068e5-8cc3-4ed1-811c-835a08eefaa0" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604354 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b5068e5-8cc3-4ed1-811c-835a08eefaa0" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604360 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604365 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604374 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604380 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604391 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604396 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604404 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604409 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604416 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604422 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604430 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604435 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604446 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604451 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604458 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604463 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604471 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604476 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604484 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604489 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604497 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604501 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604507 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604512 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604602 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604610 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604616 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604623 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b5068e5-8cc3-4ed1-811c-835a08eefaa0" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604632 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604640 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604973 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.607015 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.607122 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.611951 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.628474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.628552 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.729297 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.729391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.729393 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.743680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.924739 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:59 crc kubenswrapper[4824]: I0121 11:13:59.252083 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 21 11:13:59 crc kubenswrapper[4824]: W0121 11:13:59.256669 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbb3ac9c9_8f2c_46e7_bd3c_425f42d66a2a.slice/crio-37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7 WatchSource:0}: Error finding container 37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7: Status 404 returned error can't find the container with id 37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7 Jan 21 11:14:00 crc kubenswrapper[4824]: I0121 11:14:00.142510 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" containerID="a0e75d242bdfcf7d63ffb16f6b896e2df3292d373940f672bdb93872e312173c" exitCode=0 Jan 21 11:14:00 crc kubenswrapper[4824]: I0121 11:14:00.142543 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a","Type":"ContainerDied","Data":"a0e75d242bdfcf7d63ffb16f6b896e2df3292d373940f672bdb93872e312173c"} Jan 21 11:14:00 crc kubenswrapper[4824]: I0121 11:14:00.142701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a","Type":"ContainerStarted","Data":"37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7"} Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.394362 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.560108 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access\") pod \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.560202 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir\") pod \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.560317 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" (UID: "bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.560491 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.564494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" (UID: "bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.661434 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:02 crc kubenswrapper[4824]: I0121 11:14:02.151653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a","Type":"ContainerDied","Data":"37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7"} Jan 21 11:14:02 crc kubenswrapper[4824]: I0121 11:14:02.151687 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7" Jan 21 11:14:02 crc kubenswrapper[4824]: I0121 11:14:02.151734 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:14:03 crc kubenswrapper[4824]: I0121 11:14:03.369576 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.603320 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 21 11:14:05 crc kubenswrapper[4824]: E0121 11:14:05.603508 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" containerName="pruner" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.603519 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" containerName="pruner" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.603608 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" containerName="pruner" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.603986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.605367 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.605415 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.611040 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.699109 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.699164 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.699183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.800612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.800668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.800687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.800775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.800969 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.815664 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.915085 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:06 crc kubenswrapper[4824]: I0121 11:14:06.258768 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 21 11:14:07 crc kubenswrapper[4824]: I0121 11:14:07.175057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a83fe5bb-7790-45ea-8f0b-56799a5c4b89","Type":"ContainerStarted","Data":"8ea1c8d6b88fff12b2a363688bd7b4e3c72b3e77fe714e1ec1237fd44548b134"} Jan 21 11:14:07 crc kubenswrapper[4824]: I0121 11:14:07.175256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a83fe5bb-7790-45ea-8f0b-56799a5c4b89","Type":"ContainerStarted","Data":"939da8a2d6bcb80f9bcbb9fa2a8ace318e8eb04c7ccffba9384e787401a5fb5e"} Jan 21 11:14:07 crc kubenswrapper[4824]: I0121 11:14:07.187862 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.187849945 podStartE2EDuration="2.187849945s" podCreationTimestamp="2026-01-21 11:14:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:14:07.184792603 +0000 UTC m=+189.477821896" watchObservedRunningTime="2026-01-21 11:14:07.187849945 +0000 UTC m=+189.480879238" Jan 21 11:14:10 crc kubenswrapper[4824]: I0121 11:14:10.813577 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:14:16 crc kubenswrapper[4824]: I0121 11:14:16.065033 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:14:16 crc kubenswrapper[4824]: I0121 11:14:16.065403 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:14:35 crc kubenswrapper[4824]: I0121 11:14:35.832531 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" podUID="86b5dca2-989f-4b6b-983b-87763ce0d469" containerName="oauth-openshift" containerID="cri-o://839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f" gracePeriod=15 Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.114887 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.137631 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd"] Jan 21 11:14:36 crc kubenswrapper[4824]: E0121 11:14:36.137806 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b5dca2-989f-4b6b-983b-87763ce0d469" containerName="oauth-openshift" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.137823 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b5dca2-989f-4b6b-983b-87763ce0d469" containerName="oauth-openshift" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.137926 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b5dca2-989f-4b6b-983b-87763ce0d469" containerName="oauth-openshift" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.138250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.144610 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd"] Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200774 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200809 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200829 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhtz6\" (UniqueName: \"kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200875 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200915 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200941 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200978 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200996 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201017 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201049 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201075 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201098 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201227 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2322c755-9e87-4129-92f7-f525fbf333b4-audit-dir\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201284 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-audit-policies\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201321 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201340 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201362 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52gcl\" (UniqueName: \"kubernetes.io/projected/2322c755-9e87-4129-92f7-f525fbf333b4-kube-api-access-52gcl\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201397 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201418 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-login\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201463 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-error\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-session\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201574 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.202020 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.202111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.202108 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.202300 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.208986 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.209003 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6" (OuterVolumeSpecName: "kube-api-access-bhtz6") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "kube-api-access-bhtz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.209241 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.209494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.209664 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.209895 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.210116 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.210227 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.210417 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.282622 4824 generic.go:334] "Generic (PLEG): container finished" podID="86b5dca2-989f-4b6b-983b-87763ce0d469" containerID="839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f" exitCode=0 Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.282646 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.282659 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" event={"ID":"86b5dca2-989f-4b6b-983b-87763ce0d469","Type":"ContainerDied","Data":"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f"} Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.282709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" event={"ID":"86b5dca2-989f-4b6b-983b-87763ce0d469","Type":"ContainerDied","Data":"046a7d0640fbf37bae3b158d9284d146391f1ce208f167ac80fea14196ea051f"} Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.282725 4824 scope.go:117] "RemoveContainer" containerID="839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.295916 4824 scope.go:117] "RemoveContainer" containerID="839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f" Jan 21 11:14:36 crc kubenswrapper[4824]: E0121 11:14:36.296227 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f\": container with ID starting with 839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f not found: ID does not exist" containerID="839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.296255 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f"} err="failed to get container status \"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f\": rpc error: code = NotFound desc = could not find container \"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f\": container with ID starting with 839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f not found: ID does not exist" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.302871 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-error\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303446 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303490 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-session\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303573 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2322c755-9e87-4129-92f7-f525fbf333b4-audit-dir\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-audit-policies\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303682 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52gcl\" (UniqueName: \"kubernetes.io/projected/2322c755-9e87-4129-92f7-f525fbf333b4-kube-api-access-52gcl\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303730 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-login\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303781 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303790 4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303802 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303810 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303840 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303850 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303859 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303867 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303876 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303885 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303893 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303903 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303911 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhtz6\" (UniqueName: \"kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303920 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.304423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.304927 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2322c755-9e87-4129-92f7-f525fbf333b4-audit-dir\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.305205 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.305712 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-audit-policies\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.306209 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.306244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-session\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.306347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.306422 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-error\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.306729 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.307217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.307672 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.307859 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.308616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.308816 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-login\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.316562 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52gcl\" (UniqueName: \"kubernetes.io/projected/2322c755-9e87-4129-92f7-f525fbf333b4-kube-api-access-52gcl\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.448681 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.772423 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd"] Jan 21 11:14:37 crc kubenswrapper[4824]: I0121 11:14:37.291082 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" event={"ID":"2322c755-9e87-4129-92f7-f525fbf333b4","Type":"ContainerStarted","Data":"8f3d9e6c373bc746bb74551e87be2165a2396d15a6ad7f180620c55bda160f7e"} Jan 21 11:14:37 crc kubenswrapper[4824]: I0121 11:14:37.292185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" event={"ID":"2322c755-9e87-4129-92f7-f525fbf333b4","Type":"ContainerStarted","Data":"cd68f36f8eb5256d8f2edc35114238f1f3b7c75238f18b05388f0d06cac882df"} Jan 21 11:14:37 crc kubenswrapper[4824]: I0121 11:14:37.292465 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:37 crc kubenswrapper[4824]: I0121 11:14:37.296225 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:37 crc kubenswrapper[4824]: I0121 11:14:37.308157 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" podStartSLOduration=27.308143671 podStartE2EDuration="27.308143671s" podCreationTimestamp="2026-01-21 11:14:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:14:37.306332502 +0000 UTC m=+219.599361814" watchObservedRunningTime="2026-01-21 11:14:37.308143671 +0000 UTC m=+219.601172963" Jan 21 11:14:38 crc kubenswrapper[4824]: I0121 11:14:38.053367 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86b5dca2-989f-4b6b-983b-87763ce0d469" path="/var/lib/kubelet/pods/86b5dca2-989f-4b6b-983b-87763ce0d469/volumes" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.892564 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894169 4824 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894314 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894525 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431" gracePeriod=15 Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894685 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f" gracePeriod=15 Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894891 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233" gracePeriod=15 Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894899 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f" gracePeriod=15 Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894870 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08" gracePeriod=15 Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894879 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895215 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895230 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895241 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895246 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895257 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895262 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895271 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895276 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895300 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895306 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895313 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895319 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895409 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895419 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895426 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895436 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895459 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895555 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895563 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895652 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.917930 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.978494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.978853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979147 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979264 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979403 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979563 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979681 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080781 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080806 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080865 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080892 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080902 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080939 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.081051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.081097 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.215816 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: E0121 11:14:44.230697 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 192.168.26.120:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188cbabf404def34 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-21 11:14:44.230131508 +0000 UTC m=+226.523160800,LastTimestamp:2026-01-21 11:14:44.230131508 +0000 UTC m=+226.523160800,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.319669 4824 generic.go:334] "Generic (PLEG): container finished" podID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" containerID="8ea1c8d6b88fff12b2a363688bd7b4e3c72b3e77fe714e1ec1237fd44548b134" exitCode=0 Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.319735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a83fe5bb-7790-45ea-8f0b-56799a5c4b89","Type":"ContainerDied","Data":"8ea1c8d6b88fff12b2a363688bd7b4e3c72b3e77fe714e1ec1237fd44548b134"} Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.320261 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.320506 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.321015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"0dfee8c9edc60f4beca30cb6369cc26121c4a22045aa85ba62e4dcbcb9455144"} Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.322790 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.323631 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.324104 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f" exitCode=0 Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.324132 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233" exitCode=0 Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.324140 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08" exitCode=0 Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.324147 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f" exitCode=2 Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.324174 4824 scope.go:117] "RemoveContainer" containerID="af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.330618 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.336070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022"} Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.336589 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.336782 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.507559 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.508211 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.508509 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698256 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir\") pod \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698285 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock\") pod \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access\") pod \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock" (OuterVolumeSpecName: "var-lock") pod "a83fe5bb-7790-45ea-8f0b-56799a5c4b89" (UID: "a83fe5bb-7790-45ea-8f0b-56799a5c4b89"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a83fe5bb-7790-45ea-8f0b-56799a5c4b89" (UID: "a83fe5bb-7790-45ea-8f0b-56799a5c4b89"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698540 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698559 4824 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.702096 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a83fe5bb-7790-45ea-8f0b-56799a5c4b89" (UID: "a83fe5bb-7790-45ea-8f0b-56799a5c4b89"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.800018 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.065641 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.066027 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.066069 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.066596 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.066651 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153" gracePeriod=600 Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.166113 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.166706 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.167049 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.167258 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.167502 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305382 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305499 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305615 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305632 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305653 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305788 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305855 4824 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305866 4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305873 4824 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.342277 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153" exitCode=0 Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.342338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153"} Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.342363 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a"} Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.343093 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.343407 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.343611 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.343798 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.344002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a83fe5bb-7790-45ea-8f0b-56799a5c4b89","Type":"ContainerDied","Data":"939da8a2d6bcb80f9bcbb9fa2a8ace318e8eb04c7ccffba9384e787401a5fb5e"} Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.344025 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="939da8a2d6bcb80f9bcbb9fa2a8ace318e8eb04c7ccffba9384e787401a5fb5e" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.344057 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347272 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347594 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431" exitCode=0 Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347633 4824 scope.go:117] "RemoveContainer" containerID="3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347581 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347659 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.348046 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.349166 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.357462 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.357746 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.358121 4824 scope.go:117] "RemoveContainer" containerID="48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.358129 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.358306 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.367989 4824 scope.go:117] "RemoveContainer" containerID="d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.393224 4824 scope.go:117] "RemoveContainer" containerID="61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.402052 4824 scope.go:117] "RemoveContainer" containerID="a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.411972 4824 scope.go:117] "RemoveContainer" containerID="a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.428380 4824 scope.go:117] "RemoveContainer" containerID="3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.428900 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\": container with ID starting with 3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f not found: ID does not exist" containerID="3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.428931 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f"} err="failed to get container status \"3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\": rpc error: code = NotFound desc = could not find container \"3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\": container with ID starting with 3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.428951 4824 scope.go:117] "RemoveContainer" containerID="48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.429163 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\": container with ID starting with 48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233 not found: ID does not exist" containerID="48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429195 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233"} err="failed to get container status \"48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\": rpc error: code = NotFound desc = could not find container \"48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\": container with ID starting with 48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233 not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429207 4824 scope.go:117] "RemoveContainer" containerID="d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.429422 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\": container with ID starting with d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08 not found: ID does not exist" containerID="d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429464 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08"} err="failed to get container status \"d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\": rpc error: code = NotFound desc = could not find container \"d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\": container with ID starting with d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08 not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429478 4824 scope.go:117] "RemoveContainer" containerID="61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.429641 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\": container with ID starting with 61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f not found: ID does not exist" containerID="61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429660 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f"} err="failed to get container status \"61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\": rpc error: code = NotFound desc = could not find container \"61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\": container with ID starting with 61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429671 4824 scope.go:117] "RemoveContainer" containerID="a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.429842 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\": container with ID starting with a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431 not found: ID does not exist" containerID="a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429862 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431"} err="failed to get container status \"a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\": rpc error: code = NotFound desc = could not find container \"a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\": container with ID starting with a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431 not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429873 4824 scope.go:117] "RemoveContainer" containerID="a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.430043 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\": container with ID starting with a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc not found: ID does not exist" containerID="a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.430061 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc"} err="failed to get container status \"a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\": rpc error: code = NotFound desc = could not find container \"a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\": container with ID starting with a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.999236 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 192.168.26.120:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188cbabf404def34 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-21 11:14:44.230131508 +0000 UTC m=+226.523160800,LastTimestamp:2026-01-21 11:14:44.230131508 +0000 UTC m=+226.523160800,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 21 11:14:48 crc kubenswrapper[4824]: I0121 11:14:48.056550 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:48 crc kubenswrapper[4824]: I0121 11:14:48.057002 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 21 11:14:48 crc kubenswrapper[4824]: I0121 11:14:48.057140 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:48 crc kubenswrapper[4824]: I0121 11:14:48.057416 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:48 crc kubenswrapper[4824]: I0121 11:14:48.057723 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.552239 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:14:51Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:14:51Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:14:51Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:14:51Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.553320 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.553602 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.553797 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.554135 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.554158 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.285305 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.285662 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.285870 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.286102 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.286349 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:52 crc kubenswrapper[4824]: I0121 11:14:52.286377 4824 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.286606 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="200ms" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.487635 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="400ms" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.889042 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="800ms" Jan 21 11:14:53 crc kubenswrapper[4824]: E0121 11:14:53.690217 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="1.6s" Jan 21 11:14:54 crc kubenswrapper[4824]: E0121 11:14:54.084837 4824 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 192.168.26.120:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" volumeName="registry-storage" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.048671 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.049426 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.049823 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.050066 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.059477 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.059503 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:55 crc kubenswrapper[4824]: E0121 11:14:55.059763 4824 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.060137 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:55 crc kubenswrapper[4824]: W0121 11:14:55.074402 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-25e32e75e80b54c012f4f719d3849a2eed22b09fc73b90c2e946e0de64b5a368 WatchSource:0}: Error finding container 25e32e75e80b54c012f4f719d3849a2eed22b09fc73b90c2e946e0de64b5a368: Status 404 returned error can't find the container with id 25e32e75e80b54c012f4f719d3849a2eed22b09fc73b90c2e946e0de64b5a368 Jan 21 11:14:55 crc kubenswrapper[4824]: E0121 11:14:55.290822 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="3.2s" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381230 4824 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="39ba2e9f1acb7a8d89c60567635c87d812b096ba076466971d576e55c18f1b39" exitCode=0 Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381270 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"39ba2e9f1acb7a8d89c60567635c87d812b096ba076466971d576e55c18f1b39"} Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"25e32e75e80b54c012f4f719d3849a2eed22b09fc73b90c2e946e0de64b5a368"} Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381507 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381524 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381764 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:55 crc kubenswrapper[4824]: E0121 11:14:55.381762 4824 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381999 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.382218 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a593ed3c66b1442fd2eeb7e1fbafe41c24a6922b534938a59f914e5e6f561f51"} Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8e1c447530e12d995e9dfbdd4718c96d63cb8891dab598bfa0d0879bc4cf7f00"} Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394484 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b778529cbbba0136d4fe6b7f1c03d9e3b7f49b108c456ff9a9c6ec582a7276a7"} Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3cffa5e9fd6593f2c2eda3c289b3e16ac1a3866c802c737d57d64747711588f8"} Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394503 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b804f5b3c0033796f7109c566616c8c8430ccb2d6d9f5f537dca260fca6f87cb"} Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394675 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394744 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394758 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:58 crc kubenswrapper[4824]: I0121 11:14:58.404543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 21 11:14:58 crc kubenswrapper[4824]: I0121 11:14:58.404756 4824 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70" exitCode=1 Jan 21 11:14:58 crc kubenswrapper[4824]: I0121 11:14:58.404782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70"} Jan 21 11:14:58 crc kubenswrapper[4824]: I0121 11:14:58.405089 4824 scope.go:117] "RemoveContainer" containerID="43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70" Jan 21 11:14:59 crc kubenswrapper[4824]: I0121 11:14:59.412547 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 21 11:14:59 crc kubenswrapper[4824]: I0121 11:14:59.412802 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fead566ce315e0429506998b40292a6c707d4e650412b0c7978957b011746333"} Jan 21 11:15:00 crc kubenswrapper[4824]: I0121 11:15:00.061004 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:00 crc kubenswrapper[4824]: I0121 11:15:00.061035 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:00 crc kubenswrapper[4824]: I0121 11:15:00.065057 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:00 crc kubenswrapper[4824]: I0121 11:15:00.860275 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:15:00 crc kubenswrapper[4824]: I0121 11:15:00.863780 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:15:01 crc kubenswrapper[4824]: I0121 11:15:01.419933 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:15:01 crc kubenswrapper[4824]: I0121 11:15:01.550292 4824 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:01 crc kubenswrapper[4824]: I0121 11:15:01.598296 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3fef5d00-9d10-4b25-bca4-6482ab4c492f" Jan 21 11:15:02 crc kubenswrapper[4824]: I0121 11:15:02.423028 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:15:02 crc kubenswrapper[4824]: I0121 11:15:02.423054 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:15:02 crc kubenswrapper[4824]: I0121 11:15:02.425063 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3fef5d00-9d10-4b25-bca4-6482ab4c492f" Jan 21 11:15:08 crc kubenswrapper[4824]: I0121 11:15:08.424148 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 21 11:15:08 crc kubenswrapper[4824]: I0121 11:15:08.861616 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.088353 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.110353 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.271289 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.296717 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.332647 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.383431 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.421771 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.517897 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.754038 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.788995 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.826785 4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.862917 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 21 11:15:10 crc kubenswrapper[4824]: I0121 11:15:10.206885 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 21 11:15:10 crc kubenswrapper[4824]: I0121 11:15:10.296676 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 21 11:15:10 crc kubenswrapper[4824]: I0121 11:15:10.706471 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 21 11:15:10 crc kubenswrapper[4824]: I0121 11:15:10.846383 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 21 11:15:10 crc kubenswrapper[4824]: I0121 11:15:10.987211 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 21 11:15:11 crc kubenswrapper[4824]: I0121 11:15:11.203030 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 21 11:15:11 crc kubenswrapper[4824]: I0121 11:15:11.441187 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 21 11:15:11 crc kubenswrapper[4824]: I0121 11:15:11.496752 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 21 11:15:11 crc kubenswrapper[4824]: I0121 11:15:11.968442 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 21 11:15:12 crc kubenswrapper[4824]: I0121 11:15:12.570555 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 21 11:15:12 crc kubenswrapper[4824]: I0121 11:15:12.635523 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.142930 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.146029 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.204407 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.402258 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.471537 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.538415 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.698434 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.807767 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.941377 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.138039 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.161256 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.174452 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.281221 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.313582 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.440326 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.723166 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 21 11:15:15 crc kubenswrapper[4824]: I0121 11:15:15.405999 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 21 11:15:15 crc kubenswrapper[4824]: I0121 11:15:15.722113 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 21 11:15:15 crc kubenswrapper[4824]: I0121 11:15:15.754282 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.063754 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.114714 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.194263 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.270729 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.288735 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.412348 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.636109 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.682137 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.743942 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.832660 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.833257 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.003801 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.006805 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.061339 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.165701 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.292356 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.335771 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.372069 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.399041 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.466828 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.627782 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.635673 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.723530 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.733600 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.860717 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.952594 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.141527 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.283305 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.366524 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.369208 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.429527 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.490534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.575109 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.576739 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.701334 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.715483 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.741452 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.765931 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.796038 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.836995 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.927232 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.068057 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.111687 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.195651 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.277316 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.279188 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.294987 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.301026 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.357457 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.400529 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.417456 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.548237 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.604564 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.729577 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.767823 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.882491 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.907692 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.949904 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.962071 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.029426 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.068133 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.075603 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.093926 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.133264 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.178782 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.270048 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.286240 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.313346 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.336663 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.345823 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.373000 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.404996 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.461016 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.525112 4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.526771 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.526887 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.671061 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.714580 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.813379 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.933791 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.967106 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.038570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.101023 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.166745 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.271414 4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.272495 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=38.272480966 podStartE2EDuration="38.272480966s" podCreationTimestamp="2026-01-21 11:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:01.569012902 +0000 UTC m=+243.862042193" watchObservedRunningTime="2026-01-21 11:15:21.272480966 +0000 UTC m=+263.565510259" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.274913 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.274952 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/marketplace-operator-79b997595-xcxs9"] Jan 21 11:15:21 crc kubenswrapper[4824]: E0121 11:15:21.275159 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" containerName="installer" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.275177 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" containerName="installer" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.275270 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" containerName="installer" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.275731 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.275925 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5","openshift-marketplace/redhat-operators-hqs52","openshift-marketplace/community-operators-7q9f2","openshift-marketplace/redhat-marketplace-4w5m6","openshift-marketplace/certified-operators-b922x"] Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276108 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b922x" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="registry-server" containerID="cri-o://f4386ff52380b3c795405ecbb9387d4d9b5fca2cffa267eb0ddb32961b0a7f10" gracePeriod=30 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276114 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276280 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hqs52" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="registry-server" containerID="cri-o://b1068da5b4766f19c34535e8e58a22325fb8a2d31ce06ee547c973539c887a47" gracePeriod=30 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276409 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7q9f2" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="registry-server" containerID="cri-o://728ea008f822360aaa1493d68beaf88c96b2cdf76a7eeb252c13acb6bf91c9d5" gracePeriod=30 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276481 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" containerID="cri-o://605262189eb74a464efb5cb3518becda7f59cb81cb5a3ce736fbc1a29ea45dc4" gracePeriod=30 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276617 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4w5m6" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="registry-server" containerID="cri-o://fa2ad6f2a50b9350a988947c908be8adb082c0d1e40507b63b02b300f1699aae" gracePeriod=30 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.277698 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.278136 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.278820 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.278921 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.285645 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.294175 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.294162522 podStartE2EDuration="20.294162522s" podCreationTimestamp="2026-01-21 11:15:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:21.293228597 +0000 UTC m=+263.586257889" watchObservedRunningTime="2026-01-21 11:15:21.294162522 +0000 UTC m=+263.587191814" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.345044 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376091 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbqmv\" (UniqueName: \"kubernetes.io/projected/b4c29122-d9b2-4f85-8b29-c0e7d888667f-kube-api-access-xbqmv\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxbxn\" (UniqueName: \"kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376321 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.386204 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477128 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477189 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbqmv\" (UniqueName: \"kubernetes.io/projected/b4c29122-d9b2-4f85-8b29-c0e7d888667f-kube-api-access-xbqmv\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477223 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxbxn\" (UniqueName: \"kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477256 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.478328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.478710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.482400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.482418 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.490400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbqmv\" (UniqueName: \"kubernetes.io/projected/b4c29122-d9b2-4f85-8b29-c0e7d888667f-kube-api-access-xbqmv\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.494263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxbxn\" (UniqueName: \"kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.495979 4824 generic.go:334] "Generic (PLEG): container finished" podID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerID="b1068da5b4766f19c34535e8e58a22325fb8a2d31ce06ee547c973539c887a47" exitCode=0 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.496040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerDied","Data":"b1068da5b4766f19c34535e8e58a22325fb8a2d31ce06ee547c973539c887a47"} Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.498180 4824 generic.go:334] "Generic (PLEG): container finished" podID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerID="fa2ad6f2a50b9350a988947c908be8adb082c0d1e40507b63b02b300f1699aae" exitCode=0 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.498246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerDied","Data":"fa2ad6f2a50b9350a988947c908be8adb082c0d1e40507b63b02b300f1699aae"} Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.501816 4824 generic.go:334] "Generic (PLEG): container finished" podID="3f11911f-e365-4151-b536-3988186b8757" containerID="f4386ff52380b3c795405ecbb9387d4d9b5fca2cffa267eb0ddb32961b0a7f10" exitCode=0 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.501872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerDied","Data":"f4386ff52380b3c795405ecbb9387d4d9b5fca2cffa267eb0ddb32961b0a7f10"} Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.503828 4824 generic.go:334] "Generic (PLEG): container finished" podID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerID="728ea008f822360aaa1493d68beaf88c96b2cdf76a7eeb252c13acb6bf91c9d5" exitCode=0 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.503907 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerDied","Data":"728ea008f822360aaa1493d68beaf88c96b2cdf76a7eeb252c13acb6bf91c9d5"} Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.505062 4824 generic.go:334] "Generic (PLEG): container finished" podID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerID="605262189eb74a464efb5cb3518becda7f59cb81cb5a3ce736fbc1a29ea45dc4" exitCode=0 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.505800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" event={"ID":"14d99a1f-fca3-421a-8b47-ab4cda2f77fc","Type":"ContainerDied","Data":"605262189eb74a464efb5cb3518becda7f59cb81cb5a3ce736fbc1a29ea45dc4"} Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.557613 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5"] Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.558229 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.560146 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xcxs9"] Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.595291 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.597680 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.634450 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.680991 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.707660 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.708103 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.710728 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.715496 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.718394 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.756785 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.787081 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.821322 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.841520 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.881602 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjjkx\" (UniqueName: \"kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx\") pod \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.881643 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca\") pod \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.881667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lchhw\" (UniqueName: \"kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw\") pod \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.881694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities\") pod \"3f11911f-e365-4151-b536-3988186b8757\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882255 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content\") pod \"c3b56f2d-96c0-4597-9235-147fc4802d19\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882290 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content\") pod \"3f11911f-e365-4151-b536-3988186b8757\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities\") pod \"1b040d3d-f4a0-451a-97de-08bea41c6dce\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882342 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics\") pod \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "14d99a1f-fca3-421a-8b47-ab4cda2f77fc" (UID: "14d99a1f-fca3-421a-8b47-ab4cda2f77fc"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxt8j\" (UniqueName: \"kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j\") pod \"1b040d3d-f4a0-451a-97de-08bea41c6dce\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882429 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities\") pod \"c3b56f2d-96c0-4597-9235-147fc4802d19\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882456 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8656b\" (UniqueName: \"kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b\") pod \"c3b56f2d-96c0-4597-9235-147fc4802d19\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882490 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content\") pod \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882517 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities\") pod \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882545 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjbt6\" (UniqueName: \"kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6\") pod \"3f11911f-e365-4151-b536-3988186b8757\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882565 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content\") pod \"1b040d3d-f4a0-451a-97de-08bea41c6dce\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities" (OuterVolumeSpecName: "utilities") pod "3f11911f-e365-4151-b536-3988186b8757" (UID: "3f11911f-e365-4151-b536-3988186b8757"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882808 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882823 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.883199 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities" (OuterVolumeSpecName: "utilities") pod "1bc8c652-fb4f-48b7-86ce-c45522dfd476" (UID: "1bc8c652-fb4f-48b7-86ce-c45522dfd476"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.883259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities" (OuterVolumeSpecName: "utilities") pod "c3b56f2d-96c0-4597-9235-147fc4802d19" (UID: "c3b56f2d-96c0-4597-9235-147fc4802d19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.883835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities" (OuterVolumeSpecName: "utilities") pod "1b040d3d-f4a0-451a-97de-08bea41c6dce" (UID: "1b040d3d-f4a0-451a-97de-08bea41c6dce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.884502 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw" (OuterVolumeSpecName: "kube-api-access-lchhw") pod "1bc8c652-fb4f-48b7-86ce-c45522dfd476" (UID: "1bc8c652-fb4f-48b7-86ce-c45522dfd476"). InnerVolumeSpecName "kube-api-access-lchhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.884679 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j" (OuterVolumeSpecName: "kube-api-access-mxt8j") pod "1b040d3d-f4a0-451a-97de-08bea41c6dce" (UID: "1b040d3d-f4a0-451a-97de-08bea41c6dce"). InnerVolumeSpecName "kube-api-access-mxt8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.885028 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6" (OuterVolumeSpecName: "kube-api-access-xjbt6") pod "3f11911f-e365-4151-b536-3988186b8757" (UID: "3f11911f-e365-4151-b536-3988186b8757"). InnerVolumeSpecName "kube-api-access-xjbt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.885129 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx" (OuterVolumeSpecName: "kube-api-access-kjjkx") pod "14d99a1f-fca3-421a-8b47-ab4cda2f77fc" (UID: "14d99a1f-fca3-421a-8b47-ab4cda2f77fc"). InnerVolumeSpecName "kube-api-access-kjjkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.885354 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "14d99a1f-fca3-421a-8b47-ab4cda2f77fc" (UID: "14d99a1f-fca3-421a-8b47-ab4cda2f77fc"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.885494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b" (OuterVolumeSpecName: "kube-api-access-8656b") pod "c3b56f2d-96c0-4597-9235-147fc4802d19" (UID: "c3b56f2d-96c0-4597-9235-147fc4802d19"). InnerVolumeSpecName "kube-api-access-8656b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.901449 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bc8c652-fb4f-48b7-86ce-c45522dfd476" (UID: "1bc8c652-fb4f-48b7-86ce-c45522dfd476"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.905383 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.924401 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f11911f-e365-4151-b536-3988186b8757" (UID: "3f11911f-e365-4151-b536-3988186b8757"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.930840 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b040d3d-f4a0-451a-97de-08bea41c6dce" (UID: "1b040d3d-f4a0-451a-97de-08bea41c6dce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.979834 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3b56f2d-96c0-4597-9235-147fc4802d19" (UID: "c3b56f2d-96c0-4597-9235-147fc4802d19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983508 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxt8j\" (UniqueName: \"kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983533 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983545 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8656b\" (UniqueName: \"kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983572 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983581 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983589 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjbt6\" (UniqueName: \"kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983597 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983605 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjjkx\" (UniqueName: \"kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983613 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lchhw\" (UniqueName: \"kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983620 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983628 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983635 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983643 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.040256 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.117406 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.166872 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.211206 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.263054 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.295195 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.343149 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.384940 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.424870 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.449075 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.487776 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.496333 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.506105 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.510722 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerDied","Data":"8980ea76a1ff2b34ee51f0465162ae2c030d8714d1e864ebb767d3d7a1da8509"} Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.510753 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.510769 4824 scope.go:117] "RemoveContainer" containerID="f4386ff52380b3c795405ecbb9387d4d9b5fca2cffa267eb0ddb32961b0a7f10" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.512801 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerDied","Data":"58f9378707e041408b55a43bb2ea6d9092a260cf0c9b8d81dc167499cf1c4702"} Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.512817 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.514004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" event={"ID":"14d99a1f-fca3-421a-8b47-ab4cda2f77fc","Type":"ContainerDied","Data":"3217737690e95d2837429663404e46050468bcb19f6fbf887ccc71adb8493eef"} Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.514018 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.516594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerDied","Data":"331b281d41eb6e081cb5f5178cf5cdf5ccdc6fe5086d467942c51189e3b27cd1"} Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.516774 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.518831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerDied","Data":"1717bf3078cfbe6fee3800d069a35c471757cb03a1eddb1d8ff1e20167f89c9e"} Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.518986 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.522480 4824 scope.go:117] "RemoveContainer" containerID="bc5314fd2d078e4f04072ac4dda373754c451434b3d4dce44df1b40cbeecd1ba" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.525014 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b922x"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.528098 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b922x"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.530413 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.534935 4824 scope.go:117] "RemoveContainer" containerID="fe402f640205024f989b1231ac3c21b31c3d7e53bed03337b9aa7afe32b7848e" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.535216 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4w5m6"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.537793 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4w5m6"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.541446 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hqs52"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.544890 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hqs52"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.548821 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.548884 4824 scope.go:117] "RemoveContainer" containerID="728ea008f822360aaa1493d68beaf88c96b2cdf76a7eeb252c13acb6bf91c9d5" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.552344 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.554447 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7q9f2"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.556536 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7q9f2"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.561592 4824 scope.go:117] "RemoveContainer" containerID="9f778c1f1c308a32a3e2af072eab8265d05e159ea1a783308d30e563350f290a" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.573044 4824 scope.go:117] "RemoveContainer" containerID="ca4b551b3e03faea4b176d4748d261efd694766575aea5c759f3fe832098d94c" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.588167 4824 scope.go:117] "RemoveContainer" containerID="605262189eb74a464efb5cb3518becda7f59cb81cb5a3ce736fbc1a29ea45dc4" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.597842 4824 scope.go:117] "RemoveContainer" containerID="b1068da5b4766f19c34535e8e58a22325fb8a2d31ce06ee547c973539c887a47" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.606281 4824 scope.go:117] "RemoveContainer" containerID="2f3db218d18e31256264bc1967baec36484f60d00de32760d6c6503fe4e7b9fc" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.617856 4824 scope.go:117] "RemoveContainer" containerID="550d4757a2a4f775989081e05280aba8d60bf329b7f947d0664b2247319cf4a8" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.630233 4824 scope.go:117] "RemoveContainer" containerID="fa2ad6f2a50b9350a988947c908be8adb082c0d1e40507b63b02b300f1699aae" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.641873 4824 scope.go:117] "RemoveContainer" containerID="122ae0144dee544f8ce79c06f524513fd679f1334f79370a0f9e8b920f8d7ee4" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.657924 4824 scope.go:117] "RemoveContainer" containerID="d5b0c30b7ab307dc38d2c9ff03ad628bf5226a1bb36f64f546ca1dedef2c81fc" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.704234 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.802211 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.804359 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.876143 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.933560 4824 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.933946 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022" gracePeriod=5 Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.018476 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.024585 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.154203 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.275172 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.286295 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.319859 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.340864 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.455659 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.554579 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.589545 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.626203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.884421 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.907364 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.033290 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.034600 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.054297 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" path="/var/lib/kubelet/pods/14d99a1f-fca3-421a-8b47-ab4cda2f77fc/volumes" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.054781 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" path="/var/lib/kubelet/pods/1b040d3d-f4a0-451a-97de-08bea41c6dce/volumes" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.055390 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" path="/var/lib/kubelet/pods/1bc8c652-fb4f-48b7-86ce-c45522dfd476/volumes" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.056321 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f11911f-e365-4151-b536-3988186b8757" path="/var/lib/kubelet/pods/3f11911f-e365-4151-b536-3988186b8757/volumes" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.056872 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" path="/var/lib/kubelet/pods/c3b56f2d-96c0-4597-9235-147fc4802d19/volumes" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.161104 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.364592 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.381786 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.415754 4824 log.go:32] "RunPodSandbox from runtime service failed" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager_123c89e0-3113-4b28-9129-3848a7288052_0(d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c): error adding pod openshift-operator-lifecycle-manager_collect-profiles-29483235-4zbf5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c" Netns:"/var/run/netns/1adc7576-ade7-4a77-ada5-4aec576f0dcb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=collect-profiles-29483235-4zbf5;K8S_POD_INFRA_CONTAINER_ID=d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c;K8S_POD_UID=123c89e0-3113-4b28-9129-3848a7288052" Path:"" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5] networking: Multus: [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5/123c89e0-3113-4b28-9129-3848a7288052]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod collect-profiles-29483235-4zbf5 in out of cluster comm: pod "collect-profiles-29483235-4zbf5" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.415823 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager_123c89e0-3113-4b28-9129-3848a7288052_0(d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c): error adding pod openshift-operator-lifecycle-manager_collect-profiles-29483235-4zbf5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c" Netns:"/var/run/netns/1adc7576-ade7-4a77-ada5-4aec576f0dcb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=collect-profiles-29483235-4zbf5;K8S_POD_INFRA_CONTAINER_ID=d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c;K8S_POD_UID=123c89e0-3113-4b28-9129-3848a7288052" Path:"" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5] networking: Multus: [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5/123c89e0-3113-4b28-9129-3848a7288052]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod collect-profiles-29483235-4zbf5 in out of cluster comm: pod "collect-profiles-29483235-4zbf5" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.415843 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager_123c89e0-3113-4b28-9129-3848a7288052_0(d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c): error adding pod openshift-operator-lifecycle-manager_collect-profiles-29483235-4zbf5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c" Netns:"/var/run/netns/1adc7576-ade7-4a77-ada5-4aec576f0dcb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=collect-profiles-29483235-4zbf5;K8S_POD_INFRA_CONTAINER_ID=d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c;K8S_POD_UID=123c89e0-3113-4b28-9129-3848a7288052" Path:"" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5] networking: Multus: [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5/123c89e0-3113-4b28-9129-3848a7288052]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod collect-profiles-29483235-4zbf5 in out of cluster comm: pod "collect-profiles-29483235-4zbf5" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.415907 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager(123c89e0-3113-4b28-9129-3848a7288052)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager(123c89e0-3113-4b28-9129-3848a7288052)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager_123c89e0-3113-4b28-9129-3848a7288052_0(d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c): error adding pod openshift-operator-lifecycle-manager_collect-profiles-29483235-4zbf5 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c\\\" Netns:\\\"/var/run/netns/1adc7576-ade7-4a77-ada5-4aec576f0dcb\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=collect-profiles-29483235-4zbf5;K8S_POD_INFRA_CONTAINER_ID=d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c;K8S_POD_UID=123c89e0-3113-4b28-9129-3848a7288052\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5] networking: Multus: [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5/123c89e0-3113-4b28-9129-3848a7288052]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod collect-profiles-29483235-4zbf5 in out of cluster comm: pod \\\"collect-profiles-29483235-4zbf5\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" podUID="123c89e0-3113-4b28-9129-3848a7288052" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.528916 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.529267 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.559004 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.588501 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.614838 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.677260 4824 log.go:32] "RunPodSandbox from runtime service failed" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-xcxs9_openshift-marketplace_b4c29122-d9b2-4f85-8b29-c0e7d888667f_0(f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24): error adding pod openshift-marketplace_marketplace-operator-79b997595-xcxs9 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24" Netns:"/var/run/netns/e298ec93-b3e4-4cb3-8240-38b80e92dc97" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-xcxs9;K8S_POD_INFRA_CONTAINER_ID=f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24;K8S_POD_UID=b4c29122-d9b2-4f85-8b29-c0e7d888667f" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-xcxs9] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-xcxs9/b4c29122-d9b2-4f85-8b29-c0e7d888667f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod marketplace-operator-79b997595-xcxs9 in out of cluster comm: pod "marketplace-operator-79b997595-xcxs9" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.677309 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-xcxs9_openshift-marketplace_b4c29122-d9b2-4f85-8b29-c0e7d888667f_0(f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24): error adding pod openshift-marketplace_marketplace-operator-79b997595-xcxs9 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24" Netns:"/var/run/netns/e298ec93-b3e4-4cb3-8240-38b80e92dc97" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-xcxs9;K8S_POD_INFRA_CONTAINER_ID=f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24;K8S_POD_UID=b4c29122-d9b2-4f85-8b29-c0e7d888667f" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-xcxs9] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-xcxs9/b4c29122-d9b2-4f85-8b29-c0e7d888667f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod marketplace-operator-79b997595-xcxs9 in out of cluster comm: pod "marketplace-operator-79b997595-xcxs9" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.677325 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-xcxs9_openshift-marketplace_b4c29122-d9b2-4f85-8b29-c0e7d888667f_0(f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24): error adding pod openshift-marketplace_marketplace-operator-79b997595-xcxs9 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24" Netns:"/var/run/netns/e298ec93-b3e4-4cb3-8240-38b80e92dc97" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-xcxs9;K8S_POD_INFRA_CONTAINER_ID=f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24;K8S_POD_UID=b4c29122-d9b2-4f85-8b29-c0e7d888667f" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-xcxs9] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-xcxs9/b4c29122-d9b2-4f85-8b29-c0e7d888667f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod marketplace-operator-79b997595-xcxs9 in out of cluster comm: pod "marketplace-operator-79b997595-xcxs9" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.677369 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"marketplace-operator-79b997595-xcxs9_openshift-marketplace(b4c29122-d9b2-4f85-8b29-c0e7d888667f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"marketplace-operator-79b997595-xcxs9_openshift-marketplace(b4c29122-d9b2-4f85-8b29-c0e7d888667f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-xcxs9_openshift-marketplace_b4c29122-d9b2-4f85-8b29-c0e7d888667f_0(f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24): error adding pod openshift-marketplace_marketplace-operator-79b997595-xcxs9 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24\\\" Netns:\\\"/var/run/netns/e298ec93-b3e4-4cb3-8240-38b80e92dc97\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-xcxs9;K8S_POD_INFRA_CONTAINER_ID=f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24;K8S_POD_UID=b4c29122-d9b2-4f85-8b29-c0e7d888667f\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-xcxs9] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-xcxs9/b4c29122-d9b2-4f85-8b29-c0e7d888667f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod marketplace-operator-79b997595-xcxs9 in out of cluster comm: pod \\\"marketplace-operator-79b997595-xcxs9\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" podUID="b4c29122-d9b2-4f85-8b29-c0e7d888667f" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.837421 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.884749 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5"] Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.965758 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.125553 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.211943 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.310823 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.312306 4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.353592 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.357187 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.533824 4824 generic.go:334] "Generic (PLEG): container finished" podID="123c89e0-3113-4b28-9129-3848a7288052" containerID="1c72418277c3fff2993fb8936f438b1d2049c9f13bdefd24ee7c2900f6904eca" exitCode=0 Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.533862 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" event={"ID":"123c89e0-3113-4b28-9129-3848a7288052","Type":"ContainerDied","Data":"1c72418277c3fff2993fb8936f438b1d2049c9f13bdefd24ee7c2900f6904eca"} Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.533892 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.533902 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" event={"ID":"123c89e0-3113-4b28-9129-3848a7288052","Type":"ContainerStarted","Data":"100dbb4b83b5c4af8945cd603c66cd0fed7d5003f712c7ad794dd486ee3cadf0"} Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.534298 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.543736 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.677615 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.692944 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.802231 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.897180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.906235 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xcxs9"] Jan 21 11:15:25 crc kubenswrapper[4824]: W0121 11:15:25.912529 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4c29122_d9b2_4f85_8b29_c0e7d888667f.slice/crio-432aa6881d174864fae30b87af6d2da57882f48809f9ba7358a64d46d8bcfa8b WatchSource:0}: Error finding container 432aa6881d174864fae30b87af6d2da57882f48809f9ba7358a64d46d8bcfa8b: Status 404 returned error can't find the container with id 432aa6881d174864fae30b87af6d2da57882f48809f9ba7358a64d46d8bcfa8b Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.931235 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.938648 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.024540 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.270326 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.300168 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.349663 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.364679 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.414027 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.538295 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.538873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" event={"ID":"b4c29122-d9b2-4f85-8b29-c0e7d888667f","Type":"ContainerStarted","Data":"55b402fb3c0dcd6e053dd4d185841bbc3a89f80a565e0957c47691ca0a96a966"} Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.538931 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" event={"ID":"b4c29122-d9b2-4f85-8b29-c0e7d888667f","Type":"ContainerStarted","Data":"432aa6881d174864fae30b87af6d2da57882f48809f9ba7358a64d46d8bcfa8b"} Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.546980 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.702420 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.712766 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.728022 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.741637 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" podStartSLOduration=11.741620253 podStartE2EDuration="11.741620253s" podCreationTimestamp="2026-01-21 11:15:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:26.559199354 +0000 UTC m=+268.852228646" watchObservedRunningTime="2026-01-21 11:15:26.741620253 +0000 UTC m=+269.034649535" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.770971 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.794210 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.839336 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume\") pod \"123c89e0-3113-4b28-9129-3848a7288052\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.839368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume\") pod \"123c89e0-3113-4b28-9129-3848a7288052\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.839403 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxbxn\" (UniqueName: \"kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn\") pod \"123c89e0-3113-4b28-9129-3848a7288052\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.839892 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume" (OuterVolumeSpecName: "config-volume") pod "123c89e0-3113-4b28-9129-3848a7288052" (UID: "123c89e0-3113-4b28-9129-3848a7288052"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.847045 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "123c89e0-3113-4b28-9129-3848a7288052" (UID: "123c89e0-3113-4b28-9129-3848a7288052"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.847079 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn" (OuterVolumeSpecName: "kube-api-access-fxbxn") pod "123c89e0-3113-4b28-9129-3848a7288052" (UID: "123c89e0-3113-4b28-9129-3848a7288052"). InnerVolumeSpecName "kube-api-access-fxbxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.858546 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.890748 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.938845 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.940227 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.940246 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.940255 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxbxn\" (UniqueName: \"kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.946587 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.221499 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.246519 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.325177 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.325804 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.459552 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.479021 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.506305 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.545332 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.545346 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" event={"ID":"123c89e0-3113-4b28-9129-3848a7288052","Type":"ContainerDied","Data":"100dbb4b83b5c4af8945cd603c66cd0fed7d5003f712c7ad794dd486ee3cadf0"} Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.545399 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="100dbb4b83b5c4af8945cd603c66cd0fed7d5003f712c7ad794dd486ee3cadf0" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.545431 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.548338 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.828563 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.855235 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.894845 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.029895 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.030466 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.053751 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.061654 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.061679 4824 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="8bd8c0d7-b03c-4413-813b-aae449f2f506" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.064381 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.064407 4824 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="8bd8c0d7-b03c-4413-813b-aae449f2f506" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.138535 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.140479 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153629 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153682 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153715 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153755 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153802 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153852 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.154101 4824 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.154262 4824 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.154422 4824 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.154450 4824 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.159668 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.212651 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.254948 4824 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.548221 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.548457 4824 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022" exitCode=137 Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.548519 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.548524 4824 scope.go:117] "RemoveContainer" containerID="95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.552126 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.569008 4824 scope.go:117] "RemoveContainer" containerID="95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022" Jan 21 11:15:28 crc kubenswrapper[4824]: E0121 11:15:28.569571 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022\": container with ID starting with 95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022 not found: ID does not exist" containerID="95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.569608 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022"} err="failed to get container status \"95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022\": rpc error: code = NotFound desc = could not find container \"95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022\": container with ID starting with 95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022 not found: ID does not exist" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.677859 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.760653 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.826973 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.974946 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.015232 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.124849 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.150834 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.197422 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.210630 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.284230 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.346051 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.418897 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.535944 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.737048 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.026181 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.030503 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.066436 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.229481 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.260431 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.468469 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.797448 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.189775 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.429857 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.458534 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.533060 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.771303 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.995467 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 21 11:15:32 crc kubenswrapper[4824]: I0121 11:15:32.027976 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 21 11:15:32 crc kubenswrapper[4824]: I0121 11:15:32.211152 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 11:15:47 crc kubenswrapper[4824]: I0121 11:15:47.828136 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:15:47 crc kubenswrapper[4824]: I0121 11:15:47.828668 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" podUID="d962d0be-438a-4d08-8113-525238fa2156" containerName="controller-manager" containerID="cri-o://13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428" gracePeriod=30 Jan 21 11:15:47 crc kubenswrapper[4824]: I0121 11:15:47.922393 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:15:47 crc kubenswrapper[4824]: I0121 11:15:47.922565 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerName="route-controller-manager" containerID="cri-o://e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728" gracePeriod=30 Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.097770 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.178883 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.251883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert\") pod \"d962d0be-438a-4d08-8113-525238fa2156\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ksj6\" (UniqueName: \"kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6\") pod \"51e27630-d5c1-4eed-8232-3698ab99fc19\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca\") pod \"51e27630-d5c1-4eed-8232-3698ab99fc19\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg7z5\" (UniqueName: \"kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5\") pod \"d962d0be-438a-4d08-8113-525238fa2156\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252410 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert\") pod \"51e27630-d5c1-4eed-8232-3698ab99fc19\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252453 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca\") pod \"d962d0be-438a-4d08-8113-525238fa2156\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252499 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config\") pod \"51e27630-d5c1-4eed-8232-3698ab99fc19\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252521 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles\") pod \"d962d0be-438a-4d08-8113-525238fa2156\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252544 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config\") pod \"d962d0be-438a-4d08-8113-525238fa2156\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252881 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca" (OuterVolumeSpecName: "client-ca") pod "51e27630-d5c1-4eed-8232-3698ab99fc19" (UID: "51e27630-d5c1-4eed-8232-3698ab99fc19"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.253402 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca" (OuterVolumeSpecName: "client-ca") pod "d962d0be-438a-4d08-8113-525238fa2156" (UID: "d962d0be-438a-4d08-8113-525238fa2156"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.253410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d962d0be-438a-4d08-8113-525238fa2156" (UID: "d962d0be-438a-4d08-8113-525238fa2156"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.253436 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config" (OuterVolumeSpecName: "config") pod "51e27630-d5c1-4eed-8232-3698ab99fc19" (UID: "51e27630-d5c1-4eed-8232-3698ab99fc19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.253541 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config" (OuterVolumeSpecName: "config") pod "d962d0be-438a-4d08-8113-525238fa2156" (UID: "d962d0be-438a-4d08-8113-525238fa2156"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.256782 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5" (OuterVolumeSpecName: "kube-api-access-dg7z5") pod "d962d0be-438a-4d08-8113-525238fa2156" (UID: "d962d0be-438a-4d08-8113-525238fa2156"). InnerVolumeSpecName "kube-api-access-dg7z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.256803 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6" (OuterVolumeSpecName: "kube-api-access-5ksj6") pod "51e27630-d5c1-4eed-8232-3698ab99fc19" (UID: "51e27630-d5c1-4eed-8232-3698ab99fc19"). InnerVolumeSpecName "kube-api-access-5ksj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.256789 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "51e27630-d5c1-4eed-8232-3698ab99fc19" (UID: "51e27630-d5c1-4eed-8232-3698ab99fc19"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.257112 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d962d0be-438a-4d08-8113-525238fa2156" (UID: "d962d0be-438a-4d08-8113-525238fa2156"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353659 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353695 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353708 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353716 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353726 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ksj6\" (UniqueName: \"kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353734 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353757 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg7z5\" (UniqueName: \"kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353765 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353772 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462547 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cmbbh"] Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462726 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d962d0be-438a-4d08-8113-525238fa2156" containerName="controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462738 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d962d0be-438a-4d08-8113-525238fa2156" containerName="controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462749 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462754 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462761 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462767 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462774 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462779 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462786 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462791 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462799 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462803 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462808 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462813 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462820 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462825 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462830 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462835 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462842 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462847 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462856 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="123c89e0-3113-4b28-9129-3848a7288052" containerName="collect-profiles" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462861 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="123c89e0-3113-4b28-9129-3848a7288052" containerName="collect-profiles" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462868 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462873 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462880 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerName="route-controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462884 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerName="route-controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462893 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462897 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462906 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462912 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462918 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462924 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462931 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462936 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463056 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="123c89e0-3113-4b28-9129-3848a7288052" containerName="collect-profiles" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463066 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463074 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerName="route-controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463081 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463087 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463094 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463098 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463105 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d962d0be-438a-4d08-8113-525238fa2156" containerName="controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463111 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463663 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.466212 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.469258 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cmbbh"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.625775 4824 generic.go:334] "Generic (PLEG): container finished" podID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerID="e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728" exitCode=0 Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.625835 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" event={"ID":"51e27630-d5c1-4eed-8232-3698ab99fc19","Type":"ContainerDied","Data":"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728"} Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.625840 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.625858 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" event={"ID":"51e27630-d5c1-4eed-8232-3698ab99fc19","Type":"ContainerDied","Data":"66caec7ee49f045aa9a142ffed3e9d86fa441cb0eec919f264b1bbebed641a83"} Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.625874 4824 scope.go:117] "RemoveContainer" containerID="e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.627759 4824 generic.go:334] "Generic (PLEG): container finished" podID="d962d0be-438a-4d08-8113-525238fa2156" containerID="13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428" exitCode=0 Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.627821 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" event={"ID":"d962d0be-438a-4d08-8113-525238fa2156","Type":"ContainerDied","Data":"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428"} Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.627849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" event={"ID":"d962d0be-438a-4d08-8113-525238fa2156","Type":"ContainerDied","Data":"90b18225419fa36b91ae49d12b220ac91e071e5928e03c8bbf47f6f6ddb46af0"} Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.627849 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.638318 4824 scope.go:117] "RemoveContainer" containerID="e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.638596 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728\": container with ID starting with e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728 not found: ID does not exist" containerID="e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.638624 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728"} err="failed to get container status \"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728\": rpc error: code = NotFound desc = could not find container \"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728\": container with ID starting with e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728 not found: ID does not exist" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.638644 4824 scope.go:117] "RemoveContainer" containerID="13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.648857 4824 scope.go:117] "RemoveContainer" containerID="13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.649162 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428\": container with ID starting with 13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428 not found: ID does not exist" containerID="13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.649192 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428"} err="failed to get container status \"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428\": rpc error: code = NotFound desc = could not find container \"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428\": container with ID starting with 13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428 not found: ID does not exist" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.649573 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.652446 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.655803 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.657284 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwd7d\" (UniqueName: \"kubernetes.io/projected/3447df50-807b-487c-9acf-2a2250780bdc-kube-api-access-wwd7d\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.657315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-utilities\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.657350 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-catalog-content\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.658589 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.660936 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m4srg"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.661700 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.663105 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.666567 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m4srg"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758165 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwd7d\" (UniqueName: \"kubernetes.io/projected/3447df50-807b-487c-9acf-2a2250780bdc-kube-api-access-wwd7d\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-catalog-content\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49dkh\" (UniqueName: \"kubernetes.io/projected/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-kube-api-access-49dkh\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-utilities\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-catalog-content\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758346 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-utilities\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758733 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-catalog-content\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758769 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-utilities\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.772571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwd7d\" (UniqueName: \"kubernetes.io/projected/3447df50-807b-487c-9acf-2a2250780bdc-kube-api-access-wwd7d\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.774007 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.864110 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49dkh\" (UniqueName: \"kubernetes.io/projected/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-kube-api-access-49dkh\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.864150 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-catalog-content\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.864259 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-utilities\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.864627 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-utilities\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.864721 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-catalog-content\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.878346 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49dkh\" (UniqueName: \"kubernetes.io/projected/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-kube-api-access-49dkh\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.900320 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cmbbh"] Jan 21 11:15:48 crc kubenswrapper[4824]: W0121 11:15:48.903246 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3447df50_807b_487c_9acf_2a2250780bdc.slice/crio-8181bda55ac71db98645e6125503975812b2fa3fdcd76d88764acf074b860076 WatchSource:0}: Error finding container 8181bda55ac71db98645e6125503975812b2fa3fdcd76d88764acf074b860076: Status 404 returned error can't find the container with id 8181bda55ac71db98645e6125503975812b2fa3fdcd76d88764acf074b860076 Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.913820 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.914390 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.916721 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.917050 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.917315 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.917539 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.917586 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.917944 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.922106 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.924629 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.966665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.966901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6d2t\" (UniqueName: \"kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.966933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.966993 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.967059 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.974737 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.068120 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6d2t\" (UniqueName: \"kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.068198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.068237 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.068316 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.068354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.069155 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.069500 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.069616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.072074 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.083179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6d2t\" (UniqueName: \"kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.230078 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.305799 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m4srg"] Jan 21 11:15:49 crc kubenswrapper[4824]: W0121 11:15:49.320229 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53e1053f_f40b_4afc_ad9c_b07d46fc1f94.slice/crio-311aec34001f914806b85fb01bbd4769449448d553c52662486e8aa5470b5438 WatchSource:0}: Error finding container 311aec34001f914806b85fb01bbd4769449448d553c52662486e8aa5470b5438: Status 404 returned error can't find the container with id 311aec34001f914806b85fb01bbd4769449448d553c52662486e8aa5470b5438 Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.562294 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:49 crc kubenswrapper[4824]: W0121 11:15:49.567633 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf32d32e5_b286_475a_bda8_5d79b746e576.slice/crio-41f5c0a541981a8fe31ce891bdb4387c75d383bf8c055b1b0de1ae7af35b3bcd WatchSource:0}: Error finding container 41f5c0a541981a8fe31ce891bdb4387c75d383bf8c055b1b0de1ae7af35b3bcd: Status 404 returned error can't find the container with id 41f5c0a541981a8fe31ce891bdb4387c75d383bf8c055b1b0de1ae7af35b3bcd Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.628149 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.632430 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" event={"ID":"f32d32e5-b286-475a-bda8-5d79b746e576","Type":"ContainerStarted","Data":"41f5c0a541981a8fe31ce891bdb4387c75d383bf8c055b1b0de1ae7af35b3bcd"} Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.633992 4824 generic.go:334] "Generic (PLEG): container finished" podID="53e1053f-f40b-4afc-ad9c-b07d46fc1f94" containerID="d2d9f16df9dc20450c70352d5b5dfc3d357cc5eac304802f73897c9f9f00a93c" exitCode=0 Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.634028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4srg" event={"ID":"53e1053f-f40b-4afc-ad9c-b07d46fc1f94","Type":"ContainerDied","Data":"d2d9f16df9dc20450c70352d5b5dfc3d357cc5eac304802f73897c9f9f00a93c"} Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.634065 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4srg" event={"ID":"53e1053f-f40b-4afc-ad9c-b07d46fc1f94","Type":"ContainerStarted","Data":"311aec34001f914806b85fb01bbd4769449448d553c52662486e8aa5470b5438"} Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.637854 4824 generic.go:334] "Generic (PLEG): container finished" podID="3447df50-807b-487c-9acf-2a2250780bdc" containerID="729c6703ff41d585ba645a84b7f8de78c214ea811142e302140468d9881567aa" exitCode=0 Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.637912 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmbbh" event={"ID":"3447df50-807b-487c-9acf-2a2250780bdc","Type":"ContainerDied","Data":"729c6703ff41d585ba645a84b7f8de78c214ea811142e302140468d9881567aa"} Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.637968 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmbbh" event={"ID":"3447df50-807b-487c-9acf-2a2250780bdc","Type":"ContainerStarted","Data":"8181bda55ac71db98645e6125503975812b2fa3fdcd76d88764acf074b860076"} Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.657840 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn"] Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.658430 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.661380 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.661445 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.661593 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.661686 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.661859 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.672296 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn"] Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.672396 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.718334 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn"] Jan 21 11:15:49 crc kubenswrapper[4824]: E0121 11:15:49.718645 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-5xz7t serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" podUID="66af07e9-d9e9-4584-b458-afd995ac1eb3" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.774500 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.774544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xz7t\" (UniqueName: \"kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.774680 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.774704 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.875344 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.876058 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.876119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xz7t\" (UniqueName: \"kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.876447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.877228 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.876482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.880356 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.891562 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xz7t\" (UniqueName: \"kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.055987 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" path="/var/lib/kubelet/pods/51e27630-d5c1-4eed-8232-3698ab99fc19/volumes" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.056819 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d962d0be-438a-4d08-8113-525238fa2156" path="/var/lib/kubelet/pods/d962d0be-438a-4d08-8113-525238fa2156/volumes" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.647166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmbbh" event={"ID":"3447df50-807b-487c-9acf-2a2250780bdc","Type":"ContainerStarted","Data":"affad804ab6c5ef16e0c73a329f54a91b3bdbf2150e0045b7e2a45788bbf14f0"} Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.648890 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" event={"ID":"f32d32e5-b286-475a-bda8-5d79b746e576","Type":"ContainerStarted","Data":"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422"} Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.648913 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" podUID="f32d32e5-b286-475a-bda8-5d79b746e576" containerName="controller-manager" containerID="cri-o://7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422" gracePeriod=30 Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.648948 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.651248 4824 generic.go:334] "Generic (PLEG): container finished" podID="53e1053f-f40b-4afc-ad9c-b07d46fc1f94" containerID="df6d0dcb0faad05a6d372141361f8ee5fdb7937db0906748cd5b6648219d66ba" exitCode=0 Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.651304 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.651292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4srg" event={"ID":"53e1053f-f40b-4afc-ad9c-b07d46fc1f94","Type":"ContainerDied","Data":"df6d0dcb0faad05a6d372141361f8ee5fdb7937db0906748cd5b6648219d66ba"} Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.655682 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.657300 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.679549 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" podStartSLOduration=3.679532556 podStartE2EDuration="3.679532556s" podCreationTimestamp="2026-01-21 11:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:50.679000858 +0000 UTC m=+292.972030150" watchObservedRunningTime="2026-01-21 11:15:50.679532556 +0000 UTC m=+292.972561848" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.784690 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config\") pod \"66af07e9-d9e9-4584-b458-afd995ac1eb3\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.784761 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca\") pod \"66af07e9-d9e9-4584-b458-afd995ac1eb3\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.784801 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert\") pod \"66af07e9-d9e9-4584-b458-afd995ac1eb3\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.784838 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xz7t\" (UniqueName: \"kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t\") pod \"66af07e9-d9e9-4584-b458-afd995ac1eb3\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.785254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca" (OuterVolumeSpecName: "client-ca") pod "66af07e9-d9e9-4584-b458-afd995ac1eb3" (UID: "66af07e9-d9e9-4584-b458-afd995ac1eb3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.785288 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config" (OuterVolumeSpecName: "config") pod "66af07e9-d9e9-4584-b458-afd995ac1eb3" (UID: "66af07e9-d9e9-4584-b458-afd995ac1eb3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.788477 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "66af07e9-d9e9-4584-b458-afd995ac1eb3" (UID: "66af07e9-d9e9-4584-b458-afd995ac1eb3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.789481 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t" (OuterVolumeSpecName: "kube-api-access-5xz7t") pod "66af07e9-d9e9-4584-b458-afd995ac1eb3" (UID: "66af07e9-d9e9-4584-b458-afd995ac1eb3"). InnerVolumeSpecName "kube-api-access-5xz7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.886270 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.886303 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.886313 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xz7t\" (UniqueName: \"kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.886324 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.928092 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.959036 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:15:50 crc kubenswrapper[4824]: E0121 11:15:50.959235 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32d32e5-b286-475a-bda8-5d79b746e576" containerName="controller-manager" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.959249 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32d32e5-b286-475a-bda8-5d79b746e576" containerName="controller-manager" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.959338 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f32d32e5-b286-475a-bda8-5d79b746e576" containerName="controller-manager" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.959671 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.965523 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.059435 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rgl7w"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.060267 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.061798 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.067029 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgl7w"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088499 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config\") pod \"f32d32e5-b286-475a-bda8-5d79b746e576\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088563 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert\") pod \"f32d32e5-b286-475a-bda8-5d79b746e576\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088621 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles\") pod \"f32d32e5-b286-475a-bda8-5d79b746e576\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088638 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6d2t\" (UniqueName: \"kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t\") pod \"f32d32e5-b286-475a-bda8-5d79b746e576\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088680 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca\") pod \"f32d32e5-b286-475a-bda8-5d79b746e576\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq6rh\" (UniqueName: \"kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088916 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.089311 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f32d32e5-b286-475a-bda8-5d79b746e576" (UID: "f32d32e5-b286-475a-bda8-5d79b746e576"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.089364 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config" (OuterVolumeSpecName: "config") pod "f32d32e5-b286-475a-bda8-5d79b746e576" (UID: "f32d32e5-b286-475a-bda8-5d79b746e576"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.089477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.089591 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.089609 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.090152 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca" (OuterVolumeSpecName: "client-ca") pod "f32d32e5-b286-475a-bda8-5d79b746e576" (UID: "f32d32e5-b286-475a-bda8-5d79b746e576"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.093485 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f32d32e5-b286-475a-bda8-5d79b746e576" (UID: "f32d32e5-b286-475a-bda8-5d79b746e576"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.093646 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t" (OuterVolumeSpecName: "kube-api-access-g6d2t") pod "f32d32e5-b286-475a-bda8-5d79b746e576" (UID: "f32d32e5-b286-475a-bda8-5d79b746e576"). InnerVolumeSpecName "kube-api-access-g6d2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.190657 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.190829 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq6rh\" (UniqueName: \"kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.190926 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-catalog-content\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191000 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-utilities\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191069 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191165 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms274\" (UniqueName: \"kubernetes.io/projected/c321fe0f-f966-4865-a619-6f5f26e9bb50-kube-api-access-ms274\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191201 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191355 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191372 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6d2t\" (UniqueName: \"kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191384 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191626 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.192136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.194162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.204019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq6rh\" (UniqueName: \"kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.258076 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k2t4f"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.258866 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.260532 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.265600 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2t4f"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.281707 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.291883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms274\" (UniqueName: \"kubernetes.io/projected/c321fe0f-f966-4865-a619-6f5f26e9bb50-kube-api-access-ms274\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.291987 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-catalog-content\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.292018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-utilities\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.292382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-utilities\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.292433 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-catalog-content\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.304888 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms274\" (UniqueName: \"kubernetes.io/projected/c321fe0f-f966-4865-a619-6f5f26e9bb50-kube-api-access-ms274\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.371073 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.393212 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-925vc\" (UniqueName: \"kubernetes.io/projected/46423d25-afc7-4a50-ba21-5f7c7b7012f2-kube-api-access-925vc\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.393308 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-catalog-content\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.393329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-utilities\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.413151 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.497362 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-catalog-content\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.497393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-utilities\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.497424 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-925vc\" (UniqueName: \"kubernetes.io/projected/46423d25-afc7-4a50-ba21-5f7c7b7012f2-kube-api-access-925vc\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.498043 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-catalog-content\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.498066 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-utilities\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.510340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-925vc\" (UniqueName: \"kubernetes.io/projected/46423d25-afc7-4a50-ba21-5f7c7b7012f2-kube-api-access-925vc\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.571708 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.678991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" event={"ID":"3d23b1ac-adfc-4f40-8c75-710a71784b55","Type":"ContainerStarted","Data":"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.679045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" event={"ID":"3d23b1ac-adfc-4f40-8c75-710a71784b55","Type":"ContainerStarted","Data":"3d835dd9a1485af0ae9262121062825ff2601613bd318ffc3b5c44fd8e891dae"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.679158 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.680635 4824 generic.go:334] "Generic (PLEG): container finished" podID="f32d32e5-b286-475a-bda8-5d79b746e576" containerID="7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422" exitCode=0 Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.680680 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.680705 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" event={"ID":"f32d32e5-b286-475a-bda8-5d79b746e576","Type":"ContainerDied","Data":"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.680729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" event={"ID":"f32d32e5-b286-475a-bda8-5d79b746e576","Type":"ContainerDied","Data":"41f5c0a541981a8fe31ce891bdb4387c75d383bf8c055b1b0de1ae7af35b3bcd"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.680757 4824 scope.go:117] "RemoveContainer" containerID="7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.695655 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.697122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4srg" event={"ID":"53e1053f-f40b-4afc-ad9c-b07d46fc1f94","Type":"ContainerStarted","Data":"a68964f2e1a5aa3db059a08321155ec91cdb51762a74e4dfe8cdb3b1ebb163a3"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.700711 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" podStartSLOduration=2.700699273 podStartE2EDuration="2.700699273s" podCreationTimestamp="2026-01-21 11:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:51.700125475 +0000 UTC m=+293.993154767" watchObservedRunningTime="2026-01-21 11:15:51.700699273 +0000 UTC m=+293.993728566" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.711125 4824 generic.go:334] "Generic (PLEG): container finished" podID="3447df50-807b-487c-9acf-2a2250780bdc" containerID="affad804ab6c5ef16e0c73a329f54a91b3bdbf2150e0045b7e2a45788bbf14f0" exitCode=0 Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.711195 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.712216 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmbbh" event={"ID":"3447df50-807b-487c-9acf-2a2250780bdc","Type":"ContainerDied","Data":"affad804ab6c5ef16e0c73a329f54a91b3bdbf2150e0045b7e2a45788bbf14f0"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.723414 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m4srg" podStartSLOduration=2.231438049 podStartE2EDuration="3.723398941s" podCreationTimestamp="2026-01-21 11:15:48 +0000 UTC" firstStartedPulling="2026-01-21 11:15:49.635056416 +0000 UTC m=+291.928085708" lastFinishedPulling="2026-01-21 11:15:51.127017308 +0000 UTC m=+293.420046600" observedRunningTime="2026-01-21 11:15:51.720483986 +0000 UTC m=+294.013513288" watchObservedRunningTime="2026-01-21 11:15:51.723398941 +0000 UTC m=+294.016428232" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.738174 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.746543 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.746874 4824 scope.go:117] "RemoveContainer" containerID="7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422" Jan 21 11:15:51 crc kubenswrapper[4824]: E0121 11:15:51.749703 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422\": container with ID starting with 7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422 not found: ID does not exist" containerID="7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.749733 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422"} err="failed to get container status \"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422\": rpc error: code = NotFound desc = could not find container \"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422\": container with ID starting with 7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422 not found: ID does not exist" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.765436 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgl7w"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.787541 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.790357 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.818560 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2t4f"] Jan 21 11:15:51 crc kubenswrapper[4824]: W0121 11:15:51.825571 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46423d25_afc7_4a50_ba21_5f7c7b7012f2.slice/crio-552118ccb0cef413941998045e3b97b625892a5c504a290797b4bb8b3028c183 WatchSource:0}: Error finding container 552118ccb0cef413941998045e3b97b625892a5c504a290797b4bb8b3028c183: Status 404 returned error can't find the container with id 552118ccb0cef413941998045e3b97b625892a5c504a290797b4bb8b3028c183 Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.057587 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66af07e9-d9e9-4584-b458-afd995ac1eb3" path="/var/lib/kubelet/pods/66af07e9-d9e9-4584-b458-afd995ac1eb3/volumes" Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.058640 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f32d32e5-b286-475a-bda8-5d79b746e576" path="/var/lib/kubelet/pods/f32d32e5-b286-475a-bda8-5d79b746e576/volumes" Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.718187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmbbh" event={"ID":"3447df50-807b-487c-9acf-2a2250780bdc","Type":"ContainerStarted","Data":"89eb084b536fbfaab5d4e0ef427f60e1acf890ff15c52bf33a6e6997548d197c"} Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.719934 4824 generic.go:334] "Generic (PLEG): container finished" podID="c321fe0f-f966-4865-a619-6f5f26e9bb50" containerID="0fc010d3d5d3a46de0f9ce4eaa81396f02f3a8e9ecaa5cd0de536f8051a3c583" exitCode=0 Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.719999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgl7w" event={"ID":"c321fe0f-f966-4865-a619-6f5f26e9bb50","Type":"ContainerDied","Data":"0fc010d3d5d3a46de0f9ce4eaa81396f02f3a8e9ecaa5cd0de536f8051a3c583"} Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.720049 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgl7w" event={"ID":"c321fe0f-f966-4865-a619-6f5f26e9bb50","Type":"ContainerStarted","Data":"6f5d4e99055d52c4bf5250c538705a65a0a9ac83c8e6a13826190f5232d586a2"} Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.723138 4824 generic.go:334] "Generic (PLEG): container finished" podID="46423d25-afc7-4a50-ba21-5f7c7b7012f2" containerID="726235af80b56c27b7242f3bf79d007ef44d2287ccaa0e576724a345d8d479bf" exitCode=0 Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.723243 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2t4f" event={"ID":"46423d25-afc7-4a50-ba21-5f7c7b7012f2","Type":"ContainerDied","Data":"726235af80b56c27b7242f3bf79d007ef44d2287ccaa0e576724a345d8d479bf"} Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.723334 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2t4f" event={"ID":"46423d25-afc7-4a50-ba21-5f7c7b7012f2","Type":"ContainerStarted","Data":"552118ccb0cef413941998045e3b97b625892a5c504a290797b4bb8b3028c183"} Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.733325 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cmbbh" podStartSLOduration=2.170493884 podStartE2EDuration="4.733308733s" podCreationTimestamp="2026-01-21 11:15:48 +0000 UTC" firstStartedPulling="2026-01-21 11:15:49.642003415 +0000 UTC m=+291.935032707" lastFinishedPulling="2026-01-21 11:15:52.204818264 +0000 UTC m=+294.497847556" observedRunningTime="2026-01-21 11:15:52.730278732 +0000 UTC m=+295.023308023" watchObservedRunningTime="2026-01-21 11:15:52.733308733 +0000 UTC m=+295.026338025" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.728126 4824 generic.go:334] "Generic (PLEG): container finished" podID="c321fe0f-f966-4865-a619-6f5f26e9bb50" containerID="df8c5cff86de5a10e577959873f18fe345273d2ec855f9afa2fc62cd1e0ec39c" exitCode=0 Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.728196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgl7w" event={"ID":"c321fe0f-f966-4865-a619-6f5f26e9bb50","Type":"ContainerDied","Data":"df8c5cff86de5a10e577959873f18fe345273d2ec855f9afa2fc62cd1e0ec39c"} Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.729568 4824 generic.go:334] "Generic (PLEG): container finished" podID="46423d25-afc7-4a50-ba21-5f7c7b7012f2" containerID="f168a7e84d19eb96f1cb78ac7a1a837363cfeb22a2806201cc947b1c314eda7f" exitCode=0 Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.729636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2t4f" event={"ID":"46423d25-afc7-4a50-ba21-5f7c7b7012f2","Type":"ContainerDied","Data":"f168a7e84d19eb96f1cb78ac7a1a837363cfeb22a2806201cc947b1c314eda7f"} Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.920295 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.920852 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923170 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923348 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923402 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923438 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923592 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923648 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.930089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.029433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.029649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dxsx\" (UniqueName: \"kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.029712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.029771 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.130713 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dxsx\" (UniqueName: \"kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.130766 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.130807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.130838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.131580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.131761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.135673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.143900 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dxsx\" (UniqueName: \"kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.231555 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.591157 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.736179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgl7w" event={"ID":"c321fe0f-f966-4865-a619-6f5f26e9bb50","Type":"ContainerStarted","Data":"2475e62f3fd1d88e660f6d84396426f9e55aa1f97c8d45572e9660583aaa75a9"} Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.737815 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" event={"ID":"65ad4810-1d65-4965-b8c6-273069e2020f","Type":"ContainerStarted","Data":"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a"} Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.737860 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.738340 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" event={"ID":"65ad4810-1d65-4965-b8c6-273069e2020f","Type":"ContainerStarted","Data":"50a6e7e59a40fc7ef484103f69dec6c22086ee8fe5ecb855e025964076f2c1cf"} Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.739809 4824 patch_prober.go:28] interesting pod/route-controller-manager-75ddb6d6df-pz2wn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.739846 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.740107 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2t4f" event={"ID":"46423d25-afc7-4a50-ba21-5f7c7b7012f2","Type":"ContainerStarted","Data":"d2fc14ea25cdf27fd2552309d072f2f5fbad183d12dd8087283f643af1014aac"} Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.766374 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rgl7w" podStartSLOduration=2.280903155 podStartE2EDuration="3.76635991s" podCreationTimestamp="2026-01-21 11:15:51 +0000 UTC" firstStartedPulling="2026-01-21 11:15:52.721153431 +0000 UTC m=+295.014182723" lastFinishedPulling="2026-01-21 11:15:54.206610186 +0000 UTC m=+296.499639478" observedRunningTime="2026-01-21 11:15:54.752151092 +0000 UTC m=+297.045180384" watchObservedRunningTime="2026-01-21 11:15:54.76635991 +0000 UTC m=+297.059389202" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.767588 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" podStartSLOduration=5.767581987 podStartE2EDuration="5.767581987s" podCreationTimestamp="2026-01-21 11:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:54.765658022 +0000 UTC m=+297.058687314" watchObservedRunningTime="2026-01-21 11:15:54.767581987 +0000 UTC m=+297.060611279" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.779488 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k2t4f" podStartSLOduration=2.234450271 podStartE2EDuration="3.77946183s" podCreationTimestamp="2026-01-21 11:15:51 +0000 UTC" firstStartedPulling="2026-01-21 11:15:52.72438443 +0000 UTC m=+295.017413722" lastFinishedPulling="2026-01-21 11:15:54.269395989 +0000 UTC m=+296.562425281" observedRunningTime="2026-01-21 11:15:54.777376152 +0000 UTC m=+297.070405445" watchObservedRunningTime="2026-01-21 11:15:54.77946183 +0000 UTC m=+297.072491122" Jan 21 11:15:55 crc kubenswrapper[4824]: I0121 11:15:55.747416 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:57 crc kubenswrapper[4824]: I0121 11:15:57.910646 4824 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 21 11:15:58 crc kubenswrapper[4824]: I0121 11:15:58.775136 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:58 crc kubenswrapper[4824]: I0121 11:15:58.775174 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:58 crc kubenswrapper[4824]: I0121 11:15:58.801903 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:58 crc kubenswrapper[4824]: I0121 11:15:58.975734 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:58 crc kubenswrapper[4824]: I0121 11:15:58.975770 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:59 crc kubenswrapper[4824]: I0121 11:15:59.000245 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:59 crc kubenswrapper[4824]: I0121 11:15:59.783678 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:59 crc kubenswrapper[4824]: I0121 11:15:59.788155 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.371948 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.372222 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.398955 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.572772 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.572816 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.599356 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.795551 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.798079 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:16:07 crc kubenswrapper[4824]: I0121 11:16:07.827345 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:16:07 crc kubenswrapper[4824]: I0121 11:16:07.827687 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" podUID="3d23b1ac-adfc-4f40-8c75-710a71784b55" containerName="controller-manager" containerID="cri-o://6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d" gracePeriod=30 Jan 21 11:16:07 crc kubenswrapper[4824]: I0121 11:16:07.838590 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:16:07 crc kubenswrapper[4824]: I0121 11:16:07.838761 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" containerName="route-controller-manager" containerID="cri-o://a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a" gracePeriod=30 Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.243614 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.303380 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374106 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert\") pod \"65ad4810-1d65-4965-b8c6-273069e2020f\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374190 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config\") pod \"65ad4810-1d65-4965-b8c6-273069e2020f\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374239 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca\") pod \"65ad4810-1d65-4965-b8c6-273069e2020f\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dxsx\" (UniqueName: \"kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx\") pod \"65ad4810-1d65-4965-b8c6-273069e2020f\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374832 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config" (OuterVolumeSpecName: "config") pod "65ad4810-1d65-4965-b8c6-273069e2020f" (UID: "65ad4810-1d65-4965-b8c6-273069e2020f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374973 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca" (OuterVolumeSpecName: "client-ca") pod "65ad4810-1d65-4965-b8c6-273069e2020f" (UID: "65ad4810-1d65-4965-b8c6-273069e2020f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.378633 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "65ad4810-1d65-4965-b8c6-273069e2020f" (UID: "65ad4810-1d65-4965-b8c6-273069e2020f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.378766 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx" (OuterVolumeSpecName: "kube-api-access-7dxsx") pod "65ad4810-1d65-4965-b8c6-273069e2020f" (UID: "65ad4810-1d65-4965-b8c6-273069e2020f"). InnerVolumeSpecName "kube-api-access-7dxsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475173 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config\") pod \"3d23b1ac-adfc-4f40-8c75-710a71784b55\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475213 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq6rh\" (UniqueName: \"kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh\") pod \"3d23b1ac-adfc-4f40-8c75-710a71784b55\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca\") pod \"3d23b1ac-adfc-4f40-8c75-710a71784b55\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles\") pod \"3d23b1ac-adfc-4f40-8c75-710a71784b55\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475280 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert\") pod \"3d23b1ac-adfc-4f40-8c75-710a71784b55\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475476 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475487 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475497 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dxsx\" (UniqueName: \"kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475506 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475822 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3d23b1ac-adfc-4f40-8c75-710a71784b55" (UID: "3d23b1ac-adfc-4f40-8c75-710a71784b55"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475882 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config" (OuterVolumeSpecName: "config") pod "3d23b1ac-adfc-4f40-8c75-710a71784b55" (UID: "3d23b1ac-adfc-4f40-8c75-710a71784b55"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475895 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca" (OuterVolumeSpecName: "client-ca") pod "3d23b1ac-adfc-4f40-8c75-710a71784b55" (UID: "3d23b1ac-adfc-4f40-8c75-710a71784b55"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.477882 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3d23b1ac-adfc-4f40-8c75-710a71784b55" (UID: "3d23b1ac-adfc-4f40-8c75-710a71784b55"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.477910 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh" (OuterVolumeSpecName: "kube-api-access-jq6rh") pod "3d23b1ac-adfc-4f40-8c75-710a71784b55" (UID: "3d23b1ac-adfc-4f40-8c75-710a71784b55"). InnerVolumeSpecName "kube-api-access-jq6rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.576299 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.576327 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.576336 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.576345 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.576353 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq6rh\" (UniqueName: \"kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.796462 4824 generic.go:334] "Generic (PLEG): container finished" podID="3d23b1ac-adfc-4f40-8c75-710a71784b55" containerID="6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d" exitCode=0 Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.796500 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.796514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" event={"ID":"3d23b1ac-adfc-4f40-8c75-710a71784b55","Type":"ContainerDied","Data":"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d"} Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.796758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" event={"ID":"3d23b1ac-adfc-4f40-8c75-710a71784b55","Type":"ContainerDied","Data":"3d835dd9a1485af0ae9262121062825ff2601613bd318ffc3b5c44fd8e891dae"} Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.796776 4824 scope.go:117] "RemoveContainer" containerID="6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.798874 4824 generic.go:334] "Generic (PLEG): container finished" podID="65ad4810-1d65-4965-b8c6-273069e2020f" containerID="a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a" exitCode=0 Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.798902 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" event={"ID":"65ad4810-1d65-4965-b8c6-273069e2020f","Type":"ContainerDied","Data":"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a"} Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.798921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" event={"ID":"65ad4810-1d65-4965-b8c6-273069e2020f","Type":"ContainerDied","Data":"50a6e7e59a40fc7ef484103f69dec6c22086ee8fe5ecb855e025964076f2c1cf"} Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.798980 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.808435 4824 scope.go:117] "RemoveContainer" containerID="6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d" Jan 21 11:16:08 crc kubenswrapper[4824]: E0121 11:16:08.809473 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d\": container with ID starting with 6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d not found: ID does not exist" containerID="6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.809502 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d"} err="failed to get container status \"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d\": rpc error: code = NotFound desc = could not find container \"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d\": container with ID starting with 6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d not found: ID does not exist" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.809519 4824 scope.go:117] "RemoveContainer" containerID="a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.817210 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.822952 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.827918 4824 scope.go:117] "RemoveContainer" containerID="a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a" Jan 21 11:16:08 crc kubenswrapper[4824]: E0121 11:16:08.828489 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a\": container with ID starting with a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a not found: ID does not exist" containerID="a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.828526 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a"} err="failed to get container status \"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a\": rpc error: code = NotFound desc = could not find container \"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a\": container with ID starting with a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a not found: ID does not exist" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.829090 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.831778 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.930838 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66"] Jan 21 11:16:08 crc kubenswrapper[4824]: E0121 11:16:08.931096 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" containerName="route-controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.931109 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" containerName="route-controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: E0121 11:16:08.931122 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d23b1ac-adfc-4f40-8c75-710a71784b55" containerName="controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.931128 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d23b1ac-adfc-4f40-8c75-710a71784b55" containerName="controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.931221 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" containerName="route-controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.931231 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d23b1ac-adfc-4f40-8c75-710a71784b55" containerName="controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.931603 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.933264 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.933577 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.933687 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.933886 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.934014 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.935141 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.937445 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5975f76c65-dnkhj"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.939001 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.940223 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.940720 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.941621 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.942064 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.942321 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.942534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.942942 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5975f76c65-dnkhj"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.945623 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.945697 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5w7m\" (UniqueName: \"kubernetes.io/projected/b3bae151-0d8c-4715-a57c-77a832079b43-kube-api-access-g5w7m\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-client-ca\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981360 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-client-ca\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981401 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-config\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981416 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-proxy-ca-bundles\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-config\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee260316-013d-420c-8462-9919f6a10f12-serving-cert\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981488 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3bae151-0d8c-4715-a57c-77a832079b43-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgcf4\" (UniqueName: \"kubernetes.io/projected/ee260316-013d-420c-8462-9919f6a10f12-kube-api-access-fgcf4\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-client-ca\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082295 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-config\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082314 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-proxy-ca-bundles\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082346 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-config\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082363 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee260316-013d-420c-8462-9919f6a10f12-serving-cert\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082388 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3bae151-0d8c-4715-a57c-77a832079b43-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082414 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgcf4\" (UniqueName: \"kubernetes.io/projected/ee260316-013d-420c-8462-9919f6a10f12-kube-api-access-fgcf4\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5w7m\" (UniqueName: \"kubernetes.io/projected/b3bae151-0d8c-4715-a57c-77a832079b43-kube-api-access-g5w7m\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-client-ca\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.083275 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-client-ca\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.083488 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-client-ca\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.083695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-config\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.083738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-config\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.084494 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-proxy-ca-bundles\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.085994 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3bae151-0d8c-4715-a57c-77a832079b43-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.086662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee260316-013d-420c-8462-9919f6a10f12-serving-cert\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.094483 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgcf4\" (UniqueName: \"kubernetes.io/projected/ee260316-013d-420c-8462-9919f6a10f12-kube-api-access-fgcf4\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.094782 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5w7m\" (UniqueName: \"kubernetes.io/projected/b3bae151-0d8c-4715-a57c-77a832079b43-kube-api-access-g5w7m\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.249324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.255365 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.588410 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5975f76c65-dnkhj"] Jan 21 11:16:09 crc kubenswrapper[4824]: W0121 11:16:09.591374 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee260316_013d_420c_8462_9919f6a10f12.slice/crio-13bc2579bffa5fa9f518ccd75c1971d4a17538aa01ff7ca03fe991bc54a382bf WatchSource:0}: Error finding container 13bc2579bffa5fa9f518ccd75c1971d4a17538aa01ff7ca03fe991bc54a382bf: Status 404 returned error can't find the container with id 13bc2579bffa5fa9f518ccd75c1971d4a17538aa01ff7ca03fe991bc54a382bf Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.622120 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66"] Jan 21 11:16:09 crc kubenswrapper[4824]: W0121 11:16:09.631515 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3bae151_0d8c_4715_a57c_77a832079b43.slice/crio-35f60ff8859e4ce43d1479a7cb73566d5c30cc65d97effa7365b89c1c015acd7 WatchSource:0}: Error finding container 35f60ff8859e4ce43d1479a7cb73566d5c30cc65d97effa7365b89c1c015acd7: Status 404 returned error can't find the container with id 35f60ff8859e4ce43d1479a7cb73566d5c30cc65d97effa7365b89c1c015acd7 Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.805354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" event={"ID":"b3bae151-0d8c-4715-a57c-77a832079b43","Type":"ContainerStarted","Data":"9e977ab912f17990088ac141a2495425021fa0c75aea01ae0f65cd9dcdca4cec"} Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.805389 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" event={"ID":"b3bae151-0d8c-4715-a57c-77a832079b43","Type":"ContainerStarted","Data":"35f60ff8859e4ce43d1479a7cb73566d5c30cc65d97effa7365b89c1c015acd7"} Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.805497 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.807389 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" event={"ID":"ee260316-013d-420c-8462-9919f6a10f12","Type":"ContainerStarted","Data":"d33c304d998b36c519097b8d550522dcb828ab9bce9778ca7b1d54534b00db6f"} Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.807419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" event={"ID":"ee260316-013d-420c-8462-9919f6a10f12","Type":"ContainerStarted","Data":"13bc2579bffa5fa9f518ccd75c1971d4a17538aa01ff7ca03fe991bc54a382bf"} Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.807611 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.821262 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.831985 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" podStartSLOduration=2.831970823 podStartE2EDuration="2.831970823s" podCreationTimestamp="2026-01-21 11:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:16:09.830791758 +0000 UTC m=+312.123821050" watchObservedRunningTime="2026-01-21 11:16:09.831970823 +0000 UTC m=+312.125000114" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.833192 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" podStartSLOduration=2.833182288 podStartE2EDuration="2.833182288s" podCreationTimestamp="2026-01-21 11:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:16:09.821263922 +0000 UTC m=+312.114293214" watchObservedRunningTime="2026-01-21 11:16:09.833182288 +0000 UTC m=+312.126211580" Jan 21 11:16:10 crc kubenswrapper[4824]: I0121 11:16:10.054325 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d23b1ac-adfc-4f40-8c75-710a71784b55" path="/var/lib/kubelet/pods/3d23b1ac-adfc-4f40-8c75-710a71784b55/volumes" Jan 21 11:16:10 crc kubenswrapper[4824]: I0121 11:16:10.054971 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" path="/var/lib/kubelet/pods/65ad4810-1d65-4965-b8c6-273069e2020f/volumes" Jan 21 11:16:10 crc kubenswrapper[4824]: I0121 11:16:10.084428 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.801586 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r2kbm"] Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.802517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.818693 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r2kbm"] Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.945817 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfj4n\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-kube-api-access-mfj4n\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.945876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.945907 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-certificates\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.945933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-trusted-ca\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.946008 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-bound-sa-token\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.946148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.946173 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-tls\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.946195 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.964239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047484 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfj4n\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-kube-api-access-mfj4n\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-certificates\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-trusted-ca\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047596 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-bound-sa-token\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047665 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047682 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-tls\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.048196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.048650 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-certificates\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.048649 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-trusted-ca\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.052576 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.052602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-tls\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.066639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfj4n\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-kube-api-access-mfj4n\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.072224 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-bound-sa-token\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.116204 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.460776 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r2kbm"] Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.829372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" event={"ID":"9cd26320-635a-4df5-ba9e-3a8e45d80c8b","Type":"ContainerStarted","Data":"7f93526973f872e8f6481d2dcb95cd785ca23649c3ced89b7fb423c2be014890"} Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.829547 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" event={"ID":"9cd26320-635a-4df5-ba9e-3a8e45d80c8b","Type":"ContainerStarted","Data":"4ff9af531299c805524fe70e5d120d9b9fd04dd2983e7f0d6803f2117eb65f26"} Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.829561 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.842569 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" podStartSLOduration=1.84255385 podStartE2EDuration="1.84255385s" podCreationTimestamp="2026-01-21 11:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:16:15.84085029 +0000 UTC m=+318.133879582" watchObservedRunningTime="2026-01-21 11:16:15.84255385 +0000 UTC m=+318.135583142" Jan 21 11:16:35 crc kubenswrapper[4824]: I0121 11:16:35.120131 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:35 crc kubenswrapper[4824]: I0121 11:16:35.151163 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:16:46 crc kubenswrapper[4824]: I0121 11:16:46.065219 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:16:46 crc kubenswrapper[4824]: I0121 11:16:46.065592 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.174775 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" podUID="94e71514-0af6-4254-8815-c933ade6c9da" containerName="registry" containerID="cri-o://6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf" gracePeriod=30 Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.472688 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552763 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552831 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552858 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552885 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552905 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.553740 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.554091 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcznm\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.553768 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.553809 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.554419 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.554436 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.563248 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.563264 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.563534 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm" (OuterVolumeSpecName: "kube-api-access-zcznm") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "kube-api-access-zcznm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.563930 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.564432 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.565977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.655810 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcznm\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.655837 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.655847 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.655855 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.655863 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.000931 4824 generic.go:334] "Generic (PLEG): container finished" podID="94e71514-0af6-4254-8815-c933ade6c9da" containerID="6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf" exitCode=0 Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.000989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" event={"ID":"94e71514-0af6-4254-8815-c933ade6c9da","Type":"ContainerDied","Data":"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf"} Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.001040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" event={"ID":"94e71514-0af6-4254-8815-c933ade6c9da","Type":"ContainerDied","Data":"7d8d535dae1a2ab198548905f9a9571895c25a2b895cfb327e93cc9833852054"} Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.001060 4824 scope.go:117] "RemoveContainer" containerID="6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf" Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.002496 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.012376 4824 scope.go:117] "RemoveContainer" containerID="6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf" Jan 21 11:17:01 crc kubenswrapper[4824]: E0121 11:17:01.012641 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf\": container with ID starting with 6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf not found: ID does not exist" containerID="6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf" Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.012669 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf"} err="failed to get container status \"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf\": rpc error: code = NotFound desc = could not find container \"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf\": container with ID starting with 6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf not found: ID does not exist" Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.021267 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.023240 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:17:02 crc kubenswrapper[4824]: I0121 11:17:02.053934 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94e71514-0af6-4254-8815-c933ade6c9da" path="/var/lib/kubelet/pods/94e71514-0af6-4254-8815-c933ade6c9da/volumes" Jan 21 11:17:16 crc kubenswrapper[4824]: I0121 11:17:16.065302 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:17:16 crc kubenswrapper[4824]: I0121 11:17:16.065673 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:17:46 crc kubenswrapper[4824]: I0121 11:17:46.065178 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:17:46 crc kubenswrapper[4824]: I0121 11:17:46.065480 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:17:46 crc kubenswrapper[4824]: I0121 11:17:46.065528 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:17:46 crc kubenswrapper[4824]: I0121 11:17:46.066001 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:17:46 crc kubenswrapper[4824]: I0121 11:17:46.066065 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a" gracePeriod=600 Jan 21 11:17:47 crc kubenswrapper[4824]: I0121 11:17:47.159525 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a" exitCode=0 Jan 21 11:17:47 crc kubenswrapper[4824]: I0121 11:17:47.159599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a"} Jan 21 11:17:47 crc kubenswrapper[4824]: I0121 11:17:47.159756 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509"} Jan 21 11:17:47 crc kubenswrapper[4824]: I0121 11:17:47.159777 4824 scope.go:117] "RemoveContainer" containerID="2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153" Jan 21 11:19:46 crc kubenswrapper[4824]: I0121 11:19:46.064988 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:19:46 crc kubenswrapper[4824]: I0121 11:19:46.065331 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:19:58 crc kubenswrapper[4824]: I0121 11:19:58.162712 4824 scope.go:117] "RemoveContainer" containerID="e964ec9dc2f3dad14f23b05ef15becda9a72c91eb5ec8bd728789277a97681be" Jan 21 11:19:58 crc kubenswrapper[4824]: I0121 11:19:58.177695 4824 scope.go:117] "RemoveContainer" containerID="812b638a67ec88d3b10766fb23536e55485cc458e7ca4b815bcac4f9331b7868" Jan 21 11:19:58 crc kubenswrapper[4824]: I0121 11:19:58.189515 4824 scope.go:117] "RemoveContainer" containerID="75d2b0ba70b6aa340dd2f71ac90a2f4234cfa6ec869faff5f999306922a53203" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.146348 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp"] Jan 21 11:20:15 crc kubenswrapper[4824]: E0121 11:20:15.146929 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e71514-0af6-4254-8815-c933ade6c9da" containerName="registry" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.146943 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e71514-0af6-4254-8815-c933ade6c9da" containerName="registry" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.147101 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="94e71514-0af6-4254-8815-c933ade6c9da" containerName="registry" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.147454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.148860 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-qds56" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.149052 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.149542 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.151731 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-qhlzr"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.152315 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-qhlzr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.153392 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-mwbvc" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.158893 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.162347 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-r6qj7"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.162923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.164219 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-gn9gr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.165131 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-qhlzr"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.199718 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-r6qj7"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.341172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp54q\" (UniqueName: \"kubernetes.io/projected/b12640f9-bf75-4056-9c20-a75d63a0e177-kube-api-access-qp54q\") pod \"cert-manager-webhook-687f57d79b-r6qj7\" (UID: \"b12640f9-bf75-4056-9c20-a75d63a0e177\") " pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.341236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c7pg\" (UniqueName: \"kubernetes.io/projected/b9405067-2dfa-46f9-9471-fe456b3ca5aa-kube-api-access-6c7pg\") pod \"cert-manager-858654f9db-qhlzr\" (UID: \"b9405067-2dfa-46f9-9471-fe456b3ca5aa\") " pod="cert-manager/cert-manager-858654f9db-qhlzr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.341336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kvbn\" (UniqueName: \"kubernetes.io/projected/d783bb77-baab-4e1c-99e9-ddba7ef78fc2-kube-api-access-2kvbn\") pod \"cert-manager-cainjector-cf98fcc89-xjwkp\" (UID: \"d783bb77-baab-4e1c-99e9-ddba7ef78fc2\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.442220 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kvbn\" (UniqueName: \"kubernetes.io/projected/d783bb77-baab-4e1c-99e9-ddba7ef78fc2-kube-api-access-2kvbn\") pod \"cert-manager-cainjector-cf98fcc89-xjwkp\" (UID: \"d783bb77-baab-4e1c-99e9-ddba7ef78fc2\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.442334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp54q\" (UniqueName: \"kubernetes.io/projected/b12640f9-bf75-4056-9c20-a75d63a0e177-kube-api-access-qp54q\") pod \"cert-manager-webhook-687f57d79b-r6qj7\" (UID: \"b12640f9-bf75-4056-9c20-a75d63a0e177\") " pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.442383 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c7pg\" (UniqueName: \"kubernetes.io/projected/b9405067-2dfa-46f9-9471-fe456b3ca5aa-kube-api-access-6c7pg\") pod \"cert-manager-858654f9db-qhlzr\" (UID: \"b9405067-2dfa-46f9-9471-fe456b3ca5aa\") " pod="cert-manager/cert-manager-858654f9db-qhlzr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.458654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c7pg\" (UniqueName: \"kubernetes.io/projected/b9405067-2dfa-46f9-9471-fe456b3ca5aa-kube-api-access-6c7pg\") pod \"cert-manager-858654f9db-qhlzr\" (UID: \"b9405067-2dfa-46f9-9471-fe456b3ca5aa\") " pod="cert-manager/cert-manager-858654f9db-qhlzr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.459666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp54q\" (UniqueName: \"kubernetes.io/projected/b12640f9-bf75-4056-9c20-a75d63a0e177-kube-api-access-qp54q\") pod \"cert-manager-webhook-687f57d79b-r6qj7\" (UID: \"b12640f9-bf75-4056-9c20-a75d63a0e177\") " pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.459680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kvbn\" (UniqueName: \"kubernetes.io/projected/d783bb77-baab-4e1c-99e9-ddba7ef78fc2-kube-api-access-2kvbn\") pod \"cert-manager-cainjector-cf98fcc89-xjwkp\" (UID: \"d783bb77-baab-4e1c-99e9-ddba7ef78fc2\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.461505 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.470551 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-qhlzr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.476349 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.813643 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.820250 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.852998 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-qhlzr"] Jan 21 11:20:15 crc kubenswrapper[4824]: W0121 11:20:15.858018 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9405067_2dfa_46f9_9471_fe456b3ca5aa.slice/crio-28e66bc3cbeb909b4fb3ac4e4f61ab4e329846378d4543ec2b96cd9fdfa8db70 WatchSource:0}: Error finding container 28e66bc3cbeb909b4fb3ac4e4f61ab4e329846378d4543ec2b96cd9fdfa8db70: Status 404 returned error can't find the container with id 28e66bc3cbeb909b4fb3ac4e4f61ab4e329846378d4543ec2b96cd9fdfa8db70 Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.875187 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-r6qj7"] Jan 21 11:20:15 crc kubenswrapper[4824]: W0121 11:20:15.896334 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb12640f9_bf75_4056_9c20_a75d63a0e177.slice/crio-b1517ed0a3cfaa749f6d4f7dc9db45c31ec848c0133cd19151049d63d2d45be3 WatchSource:0}: Error finding container b1517ed0a3cfaa749f6d4f7dc9db45c31ec848c0133cd19151049d63d2d45be3: Status 404 returned error can't find the container with id b1517ed0a3cfaa749f6d4f7dc9db45c31ec848c0133cd19151049d63d2d45be3 Jan 21 11:20:16 crc kubenswrapper[4824]: I0121 11:20:16.065084 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:20:16 crc kubenswrapper[4824]: I0121 11:20:16.065128 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:20:16 crc kubenswrapper[4824]: I0121 11:20:16.736034 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" event={"ID":"d783bb77-baab-4e1c-99e9-ddba7ef78fc2","Type":"ContainerStarted","Data":"75d0a83e78ed8aa21e24436721746b82415add7b87f03e4bebcd002f4587d885"} Jan 21 11:20:16 crc kubenswrapper[4824]: I0121 11:20:16.736977 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-qhlzr" event={"ID":"b9405067-2dfa-46f9-9471-fe456b3ca5aa","Type":"ContainerStarted","Data":"28e66bc3cbeb909b4fb3ac4e4f61ab4e329846378d4543ec2b96cd9fdfa8db70"} Jan 21 11:20:16 crc kubenswrapper[4824]: I0121 11:20:16.738077 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" event={"ID":"b12640f9-bf75-4056-9c20-a75d63a0e177","Type":"ContainerStarted","Data":"b1517ed0a3cfaa749f6d4f7dc9db45c31ec848c0133cd19151049d63d2d45be3"} Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.749231 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" event={"ID":"d783bb77-baab-4e1c-99e9-ddba7ef78fc2","Type":"ContainerStarted","Data":"4d041b2f34851fd9a7b84d0a992dcaddfda503a570a95e13ab3cd028a73af7b5"} Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.750605 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-qhlzr" event={"ID":"b9405067-2dfa-46f9-9471-fe456b3ca5aa","Type":"ContainerStarted","Data":"c22e6735006a7cf615e2c6c3f3835939e153c7d76b78ede8f850ddea1e043bd6"} Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.752154 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" event={"ID":"b12640f9-bf75-4056-9c20-a75d63a0e177","Type":"ContainerStarted","Data":"35dc60dcf46360dc9d6bc1df7c69987e9a40e749194976d90d26ec7dfa44a288"} Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.752277 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.775329 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" podStartSLOduration=1.00580889 podStartE2EDuration="3.775312004s" podCreationTimestamp="2026-01-21 11:20:15 +0000 UTC" firstStartedPulling="2026-01-21 11:20:15.820053051 +0000 UTC m=+558.113082344" lastFinishedPulling="2026-01-21 11:20:18.589556166 +0000 UTC m=+560.882585458" observedRunningTime="2026-01-21 11:20:18.771884718 +0000 UTC m=+561.064914010" watchObservedRunningTime="2026-01-21 11:20:18.775312004 +0000 UTC m=+561.068341296" Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.795491 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" podStartSLOduration=1.08249887 podStartE2EDuration="3.795474777s" podCreationTimestamp="2026-01-21 11:20:15 +0000 UTC" firstStartedPulling="2026-01-21 11:20:15.902882567 +0000 UTC m=+558.195911859" lastFinishedPulling="2026-01-21 11:20:18.615858464 +0000 UTC m=+560.908887766" observedRunningTime="2026-01-21 11:20:18.794617602 +0000 UTC m=+561.087646894" watchObservedRunningTime="2026-01-21 11:20:18.795474777 +0000 UTC m=+561.088504069" Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.816240 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-qhlzr" podStartSLOduration=1.085658702 podStartE2EDuration="3.816218413s" podCreationTimestamp="2026-01-21 11:20:15 +0000 UTC" firstStartedPulling="2026-01-21 11:20:15.859708225 +0000 UTC m=+558.152737517" lastFinishedPulling="2026-01-21 11:20:18.590267937 +0000 UTC m=+560.883297228" observedRunningTime="2026-01-21 11:20:18.81577043 +0000 UTC m=+561.108799723" watchObservedRunningTime="2026-01-21 11:20:18.816218413 +0000 UTC m=+561.109247706" Jan 21 11:20:25 crc kubenswrapper[4824]: I0121 11:20:25.479363 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.194441 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-htkvk"] Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.194929 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-controller" containerID="cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.194999 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.195035 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="sbdb" containerID="cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.195110 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-node" containerID="cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.194986 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="nbdb" containerID="cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.195089 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-acl-logging" containerID="cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.195068 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="northd" containerID="cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.219889 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" containerID="cri-o://99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.455485 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/3.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.457717 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovn-acl-logging/0.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.458072 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovn-controller/0.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.458361 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504329 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-r8p75"] Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504634 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504647 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504657 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504663 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504675 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504680 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504695 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="nbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504702 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="nbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504726 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kubecfg-setup" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504732 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kubecfg-setup" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504743 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-acl-logging" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504748 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-acl-logging" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504755 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-node" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504767 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-node" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504773 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="sbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504779 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="sbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504790 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504796 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504806 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-ovn-metrics" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504812 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-ovn-metrics" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504818 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="northd" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504823 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="northd" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.505644 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.505683 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.505699 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.505711 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-ovn-metrics" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506411 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-acl-logging" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506438 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="northd" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506457 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="sbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506465 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-node" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506473 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506486 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="nbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.506892 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506911 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.506920 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506925 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.507045 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.507236 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.508675 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556821 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556856 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556872 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556892 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556887 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556909 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556943 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556948 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash" (OuterVolumeSpecName: "host-slash") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556996 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket" (OuterVolumeSpecName: "log-socket") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556993 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557007 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557068 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557092 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557142 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557136 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557162 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557169 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557184 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557276 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557210 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557316 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557327 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557341 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log" (OuterVolumeSpecName: "node-log") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557346 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcf8q\" (UniqueName: \"kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557407 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557411 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557452 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557484 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557526 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557570 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557813 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557835 4824 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557846 4824 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557858 4824 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557867 4824 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557867 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557877 4824 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557888 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557899 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557907 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557916 4824 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557925 4824 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557933 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557941 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557950 4824 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557974 4824 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557983 4824 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.561303 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q" (OuterVolumeSpecName: "kube-api-access-zcf8q") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "kube-api-access-zcf8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.561313 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.567339 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-env-overrides\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-systemd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-systemd-units\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-node-log\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659491 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-var-lib-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659529 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-log-socket\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-netd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-ovn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659575 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-script-lib\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659626 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf7mn\" (UniqueName: \"kubernetes.io/projected/177dd5b5-7311-483f-8c21-0dae78cd3d09-kube-api-access-cf7mn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659642 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-kubelet\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-slash\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovn-node-metrics-cert\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659735 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-etc-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-netns\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-config\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659866 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659939 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-bin\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.660047 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.660063 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcf8q\" (UniqueName: \"kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.660074 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.660084 4824 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-bin\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761126 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-env-overrides\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761150 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-systemd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761176 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-systemd-units\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-node-log\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-bin\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-var-lib-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-var-lib-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-systemd-units\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-log-socket\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-log-socket\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-systemd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-netd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-netd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761322 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-ovn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-script-lib\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761358 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-ovn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761387 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf7mn\" (UniqueName: \"kubernetes.io/projected/177dd5b5-7311-483f-8c21-0dae78cd3d09-kube-api-access-cf7mn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761406 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761427 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-kubelet\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761444 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-slash\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovn-node-metrics-cert\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-etc-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761522 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-netns\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761538 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-config\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-kubelet\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761561 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761590 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-netns\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761608 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-etc-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761609 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-env-overrides\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-slash\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761763 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-node-log\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761941 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-script-lib\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.762235 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-config\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.764558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovn-node-metrics-cert\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.775943 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf7mn\" (UniqueName: \"kubernetes.io/projected/177dd5b5-7311-483f-8c21-0dae78cd3d09-kube-api-access-cf7mn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.783319 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/3.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.784983 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovn-acl-logging/0.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785358 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovn-controller/0.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785679 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785699 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785707 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785713 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785721 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785727 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785732 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" exitCode=143 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785737 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" exitCode=143 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785746 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785751 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785783 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785794 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785804 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785813 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785821 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785830 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785832 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785840 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785933 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785946 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785951 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785971 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785976 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785981 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785986 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786016 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786021 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786027 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786031 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786036 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786040 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786045 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786049 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786053 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786058 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786064 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786072 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786079 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786084 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786089 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786093 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786098 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786102 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786107 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786114 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786119 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786125 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"b233a1680204a0985feed02a396e93f06104160dd5b2e0d82128d39c672a1129"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786132 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786137 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786142 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786146 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786151 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786155 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786160 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786164 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786168 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786172 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787143 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/2.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787488 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/1.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787530 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" containerID="80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470" exitCode=2 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787549 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerDied","Data":"80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787563 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787930 4824 scope.go:117] "RemoveContainer" containerID="80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.788169 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x94nb_openshift-multus(6b9588d1-5876-4c2c-8de5-e5c0ad46f04e)\"" pod="openshift-multus/multus-x94nb" podUID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.797952 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.813619 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-htkvk"] Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.816022 4824 scope.go:117] "RemoveContainer" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.816492 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-htkvk"] Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.819766 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.824981 4824 scope.go:117] "RemoveContainer" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.836071 4824 scope.go:117] "RemoveContainer" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.844881 4824 scope.go:117] "RemoveContainer" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.856076 4824 scope.go:117] "RemoveContainer" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.866356 4824 scope.go:117] "RemoveContainer" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.886791 4824 scope.go:117] "RemoveContainer" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.896911 4824 scope.go:117] "RemoveContainer" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.906210 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.906519 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.906552 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} err="failed to get container status \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.906573 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.906779 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": container with ID starting with 86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483 not found: ID does not exist" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.906804 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} err="failed to get container status \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": rpc error: code = NotFound desc = could not find container \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": container with ID starting with 86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.906824 4824 scope.go:117] "RemoveContainer" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.907044 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": container with ID starting with 948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b not found: ID does not exist" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.907089 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} err="failed to get container status \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": rpc error: code = NotFound desc = could not find container \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": container with ID starting with 948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.907105 4824 scope.go:117] "RemoveContainer" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.907330 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": container with ID starting with fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c not found: ID does not exist" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.907354 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} err="failed to get container status \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": rpc error: code = NotFound desc = could not find container \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": container with ID starting with fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.907419 4824 scope.go:117] "RemoveContainer" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.907975 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": container with ID starting with ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38 not found: ID does not exist" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908001 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} err="failed to get container status \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": rpc error: code = NotFound desc = could not find container \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": container with ID starting with ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908016 4824 scope.go:117] "RemoveContainer" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.908212 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": container with ID starting with cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc not found: ID does not exist" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908258 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} err="failed to get container status \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": rpc error: code = NotFound desc = could not find container \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": container with ID starting with cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908272 4824 scope.go:117] "RemoveContainer" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.908507 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": container with ID starting with fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b not found: ID does not exist" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908528 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} err="failed to get container status \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": rpc error: code = NotFound desc = could not find container \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": container with ID starting with fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908544 4824 scope.go:117] "RemoveContainer" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.908802 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": container with ID starting with 64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3 not found: ID does not exist" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908829 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} err="failed to get container status \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": rpc error: code = NotFound desc = could not find container \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": container with ID starting with 64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908843 4824 scope.go:117] "RemoveContainer" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.909127 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": container with ID starting with 5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880 not found: ID does not exist" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909150 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} err="failed to get container status \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": rpc error: code = NotFound desc = could not find container \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": container with ID starting with 5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909167 4824 scope.go:117] "RemoveContainer" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.909389 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": container with ID starting with 82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3 not found: ID does not exist" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909408 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} err="failed to get container status \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": rpc error: code = NotFound desc = could not find container \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": container with ID starting with 82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909421 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909620 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} err="failed to get container status \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909639 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909821 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} err="failed to get container status \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": rpc error: code = NotFound desc = could not find container \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": container with ID starting with 86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909839 4824 scope.go:117] "RemoveContainer" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.910648 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} err="failed to get container status \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": rpc error: code = NotFound desc = could not find container \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": container with ID starting with 948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.910668 4824 scope.go:117] "RemoveContainer" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.910999 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} err="failed to get container status \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": rpc error: code = NotFound desc = could not find container \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": container with ID starting with fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.911021 4824 scope.go:117] "RemoveContainer" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.911339 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} err="failed to get container status \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": rpc error: code = NotFound desc = could not find container \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": container with ID starting with ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.911419 4824 scope.go:117] "RemoveContainer" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.911619 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} err="failed to get container status \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": rpc error: code = NotFound desc = could not find container \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": container with ID starting with cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.911649 4824 scope.go:117] "RemoveContainer" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912123 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} err="failed to get container status \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": rpc error: code = NotFound desc = could not find container \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": container with ID starting with fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912144 4824 scope.go:117] "RemoveContainer" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912397 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} err="failed to get container status \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": rpc error: code = NotFound desc = could not find container \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": container with ID starting with 64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912436 4824 scope.go:117] "RemoveContainer" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912839 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} err="failed to get container status \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": rpc error: code = NotFound desc = could not find container \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": container with ID starting with 5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912860 4824 scope.go:117] "RemoveContainer" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.913346 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} err="failed to get container status \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": rpc error: code = NotFound desc = could not find container \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": container with ID starting with 82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.913371 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.914151 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} err="failed to get container status \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.914175 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.914737 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} err="failed to get container status \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": rpc error: code = NotFound desc = could not find container \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": container with ID starting with 86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.914758 4824 scope.go:117] "RemoveContainer" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.915106 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} err="failed to get container status \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": rpc error: code = NotFound desc = could not find container \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": container with ID starting with 948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.915178 4824 scope.go:117] "RemoveContainer" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.915707 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} err="failed to get container status \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": rpc error: code = NotFound desc = could not find container \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": container with ID starting with fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.915732 4824 scope.go:117] "RemoveContainer" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916030 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} err="failed to get container status \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": rpc error: code = NotFound desc = could not find container \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": container with ID starting with ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916053 4824 scope.go:117] "RemoveContainer" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916226 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} err="failed to get container status \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": rpc error: code = NotFound desc = could not find container \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": container with ID starting with cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916241 4824 scope.go:117] "RemoveContainer" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916397 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} err="failed to get container status \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": rpc error: code = NotFound desc = could not find container \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": container with ID starting with fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916415 4824 scope.go:117] "RemoveContainer" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916582 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} err="failed to get container status \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": rpc error: code = NotFound desc = could not find container \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": container with ID starting with 64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916599 4824 scope.go:117] "RemoveContainer" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916819 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} err="failed to get container status \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": rpc error: code = NotFound desc = could not find container \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": container with ID starting with 5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916834 4824 scope.go:117] "RemoveContainer" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917203 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} err="failed to get container status \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": rpc error: code = NotFound desc = could not find container \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": container with ID starting with 82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917221 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917723 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} err="failed to get container status \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917738 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917920 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} err="failed to get container status \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": rpc error: code = NotFound desc = could not find container \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": container with ID starting with 86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917936 4824 scope.go:117] "RemoveContainer" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918285 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} err="failed to get container status \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": rpc error: code = NotFound desc = could not find container \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": container with ID starting with 948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918317 4824 scope.go:117] "RemoveContainer" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918538 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} err="failed to get container status \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": rpc error: code = NotFound desc = could not find container \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": container with ID starting with fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918563 4824 scope.go:117] "RemoveContainer" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918839 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} err="failed to get container status \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": rpc error: code = NotFound desc = could not find container \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": container with ID starting with ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918866 4824 scope.go:117] "RemoveContainer" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919270 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} err="failed to get container status \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": rpc error: code = NotFound desc = could not find container \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": container with ID starting with cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919291 4824 scope.go:117] "RemoveContainer" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919456 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} err="failed to get container status \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": rpc error: code = NotFound desc = could not find container \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": container with ID starting with fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919480 4824 scope.go:117] "RemoveContainer" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919764 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} err="failed to get container status \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": rpc error: code = NotFound desc = could not find container \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": container with ID starting with 64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919787 4824 scope.go:117] "RemoveContainer" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.920120 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} err="failed to get container status \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": rpc error: code = NotFound desc = could not find container \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": container with ID starting with 5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.920140 4824 scope.go:117] "RemoveContainer" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.920309 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} err="failed to get container status \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": rpc error: code = NotFound desc = could not find container \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": container with ID starting with 82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.920326 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.920520 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} err="failed to get container status \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" Jan 21 11:20:27 crc kubenswrapper[4824]: I0121 11:20:27.793797 4824 generic.go:334] "Generic (PLEG): container finished" podID="177dd5b5-7311-483f-8c21-0dae78cd3d09" containerID="dba9e5aa729335a50db7092a27a944bd16d57947741a4fe22c5ebbecd3bd67d3" exitCode=0 Jan 21 11:20:27 crc kubenswrapper[4824]: I0121 11:20:27.793847 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerDied","Data":"dba9e5aa729335a50db7092a27a944bd16d57947741a4fe22c5ebbecd3bd67d3"} Jan 21 11:20:27 crc kubenswrapper[4824]: I0121 11:20:27.794026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"2d3f3ce0ea528595b97242e1fdaded28cfa81f33c89f8948464949ba03f16b59"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.054083 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" path="/var/lib/kubelet/pods/26b1108a-4b98-4e39-a3cd-e0d055089fd8/volumes" Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"9ce84d1b9a0df30346d416335b9702c66fa67375f6060d703f03b3698a0f8673"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"7d8ed1d69da2c094a0c9e24cdf8fce33f5e3c95e0b157a8b595e4220bed4a1da"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801205 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"b5bb983c3637b349c5675ac216689b556a6378d13e14e9565faaa1a42d4f10bf"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801214 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"1b36a9038b6c5163b2384d2b4ba0cbfe0a0f50318802f964ccd26242e63c7f48"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"54d7dc9977042eacfe440060eef65173d1c15fa89af0c71f3f672e6c5a24c539"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801231 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"0ec365d42cddf75d8f253505f55690d7185fa643d4972f13b804be54193e743c"} Jan 21 11:20:30 crc kubenswrapper[4824]: I0121 11:20:30.814837 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"92696eaafafab924a846f8ec543265b14a1d4eeddd3ea442d5e797aa54f4e71b"} Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.825462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"87209c53211cdf230973b41ca812317014158ddab635b4d6b0acc166ec380ccd"} Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.825700 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.825886 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.825900 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.845445 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" podStartSLOduration=6.845431265 podStartE2EDuration="6.845431265s" podCreationTimestamp="2026-01-21 11:20:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:20:32.843111304 +0000 UTC m=+575.136140596" watchObservedRunningTime="2026-01-21 11:20:32.845431265 +0000 UTC m=+575.138460557" Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.845807 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.850436 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:40 crc kubenswrapper[4824]: I0121 11:20:40.049532 4824 scope.go:117] "RemoveContainer" containerID="80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470" Jan 21 11:20:40 crc kubenswrapper[4824]: E0121 11:20:40.049939 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x94nb_openshift-multus(6b9588d1-5876-4c2c-8de5-e5c0ad46f04e)\"" pod="openshift-multus/multus-x94nb" podUID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.064668 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.065060 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.065092 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.065519 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.065569 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509" gracePeriod=600 Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.883214 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509" exitCode=0 Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.883271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509"} Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.883520 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44"} Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.883539 4824 scope.go:117] "RemoveContainer" containerID="04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a" Jan 21 11:20:53 crc kubenswrapper[4824]: I0121 11:20:53.049401 4824 scope.go:117] "RemoveContainer" containerID="80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470" Jan 21 11:20:53 crc kubenswrapper[4824]: I0121 11:20:53.916115 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/2.log" Jan 21 11:20:53 crc kubenswrapper[4824]: I0121 11:20:53.916724 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/1.log" Jan 21 11:20:53 crc kubenswrapper[4824]: I0121 11:20:53.916865 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerStarted","Data":"29d5c1eba74f1cd555791e1c196c0fb86bfc85e8e7e102815995a7488969c8b1"} Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.524444 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4"] Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.525469 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.527033 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.532465 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4"] Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.653385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm5bq\" (UniqueName: \"kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.653489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.653528 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.754342 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.754391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.754445 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm5bq\" (UniqueName: \"kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.754877 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.754924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.768574 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm5bq\" (UniqueName: \"kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.836407 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:55 crc kubenswrapper[4824]: I0121 11:20:55.171003 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4"] Jan 21 11:20:55 crc kubenswrapper[4824]: W0121 11:20:55.174841 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2dd714a_8d35_49a1_b6c2_7bec802eed41.slice/crio-221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029 WatchSource:0}: Error finding container 221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029: Status 404 returned error can't find the container with id 221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029 Jan 21 11:20:55 crc kubenswrapper[4824]: I0121 11:20:55.933367 4824 generic.go:334] "Generic (PLEG): container finished" podID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerID="ce25b5463ed4235aca7c296c4ce401a068db4acbc17350d0ada1212aa35d509a" exitCode=0 Jan 21 11:20:55 crc kubenswrapper[4824]: I0121 11:20:55.933429 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" event={"ID":"d2dd714a-8d35-49a1-b6c2-7bec802eed41","Type":"ContainerDied","Data":"ce25b5463ed4235aca7c296c4ce401a068db4acbc17350d0ada1212aa35d509a"} Jan 21 11:20:55 crc kubenswrapper[4824]: I0121 11:20:55.934143 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" event={"ID":"d2dd714a-8d35-49a1-b6c2-7bec802eed41","Type":"ContainerStarted","Data":"221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029"} Jan 21 11:20:56 crc kubenswrapper[4824]: I0121 11:20:56.837141 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:57 crc kubenswrapper[4824]: I0121 11:20:57.943164 4824 generic.go:334] "Generic (PLEG): container finished" podID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerID="fe4aa38cf65a78901ded8e31785ce9f1dc40110d372e124bc39a42bc8328b91e" exitCode=0 Jan 21 11:20:57 crc kubenswrapper[4824]: I0121 11:20:57.943260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" event={"ID":"d2dd714a-8d35-49a1-b6c2-7bec802eed41","Type":"ContainerDied","Data":"fe4aa38cf65a78901ded8e31785ce9f1dc40110d372e124bc39a42bc8328b91e"} Jan 21 11:20:58 crc kubenswrapper[4824]: I0121 11:20:58.210078 4824 scope.go:117] "RemoveContainer" containerID="49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f" Jan 21 11:20:58 crc kubenswrapper[4824]: I0121 11:20:58.948543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/2.log" Jan 21 11:20:58 crc kubenswrapper[4824]: I0121 11:20:58.950234 4824 generic.go:334] "Generic (PLEG): container finished" podID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerID="e59780e1ea6b956c60ca193737940a3132676cbfa71031ce443e2652855dc396" exitCode=0 Jan 21 11:20:58 crc kubenswrapper[4824]: I0121 11:20:58.950264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" event={"ID":"d2dd714a-8d35-49a1-b6c2-7bec802eed41","Type":"ContainerDied","Data":"e59780e1ea6b956c60ca193737940a3132676cbfa71031ce443e2652855dc396"} Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.137337 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.313641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle\") pod \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.313691 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm5bq\" (UniqueName: \"kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq\") pod \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.313795 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util\") pod \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.314145 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle" (OuterVolumeSpecName: "bundle") pod "d2dd714a-8d35-49a1-b6c2-7bec802eed41" (UID: "d2dd714a-8d35-49a1-b6c2-7bec802eed41"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.318152 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq" (OuterVolumeSpecName: "kube-api-access-sm5bq") pod "d2dd714a-8d35-49a1-b6c2-7bec802eed41" (UID: "d2dd714a-8d35-49a1-b6c2-7bec802eed41"). InnerVolumeSpecName "kube-api-access-sm5bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.323923 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util" (OuterVolumeSpecName: "util") pod "d2dd714a-8d35-49a1-b6c2-7bec802eed41" (UID: "d2dd714a-8d35-49a1-b6c2-7bec802eed41"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.414683 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.414708 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.414718 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm5bq\" (UniqueName: \"kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.961168 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" event={"ID":"d2dd714a-8d35-49a1-b6c2-7bec802eed41","Type":"ContainerDied","Data":"221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029"} Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.961203 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.961214 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363258 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-zmc4k"] Jan 21 11:21:02 crc kubenswrapper[4824]: E0121 11:21:02.363434 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="util" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363445 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="util" Jan 21 11:21:02 crc kubenswrapper[4824]: E0121 11:21:02.363462 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="pull" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363467 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="pull" Jan 21 11:21:02 crc kubenswrapper[4824]: E0121 11:21:02.363473 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="extract" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363500 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="extract" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363617 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="extract" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363937 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.365994 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-txpxx" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.365993 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.366706 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.376100 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-zmc4k"] Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.536329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kmt6\" (UniqueName: \"kubernetes.io/projected/df60ca53-0573-4a6a-9235-86e0a1c17f91-kube-api-access-5kmt6\") pod \"nmstate-operator-646758c888-zmc4k\" (UID: \"df60ca53-0573-4a6a-9235-86e0a1c17f91\") " pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.637858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kmt6\" (UniqueName: \"kubernetes.io/projected/df60ca53-0573-4a6a-9235-86e0a1c17f91-kube-api-access-5kmt6\") pod \"nmstate-operator-646758c888-zmc4k\" (UID: \"df60ca53-0573-4a6a-9235-86e0a1c17f91\") " pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.651724 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kmt6\" (UniqueName: \"kubernetes.io/projected/df60ca53-0573-4a6a-9235-86e0a1c17f91-kube-api-access-5kmt6\") pod \"nmstate-operator-646758c888-zmc4k\" (UID: \"df60ca53-0573-4a6a-9235-86e0a1c17f91\") " pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.676127 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" Jan 21 11:21:03 crc kubenswrapper[4824]: I0121 11:21:03.044416 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-zmc4k"] Jan 21 11:21:03 crc kubenswrapper[4824]: W0121 11:21:03.049478 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf60ca53_0573_4a6a_9235_86e0a1c17f91.slice/crio-2ebdc5a7fcbbac549ba525a48829fc07aa41ebd21fa133f17578560b4ad02d03 WatchSource:0}: Error finding container 2ebdc5a7fcbbac549ba525a48829fc07aa41ebd21fa133f17578560b4ad02d03: Status 404 returned error can't find the container with id 2ebdc5a7fcbbac549ba525a48829fc07aa41ebd21fa133f17578560b4ad02d03 Jan 21 11:21:03 crc kubenswrapper[4824]: I0121 11:21:03.974338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" event={"ID":"df60ca53-0573-4a6a-9235-86e0a1c17f91","Type":"ContainerStarted","Data":"2ebdc5a7fcbbac549ba525a48829fc07aa41ebd21fa133f17578560b4ad02d03"} Jan 21 11:21:04 crc kubenswrapper[4824]: I0121 11:21:04.980027 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" event={"ID":"df60ca53-0573-4a6a-9235-86e0a1c17f91","Type":"ContainerStarted","Data":"2bf180cba71ee6b1e9512a8e513b6e4e50a4d442dae7a116fb2209f499f0cbbf"} Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.735460 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" podStartSLOduration=2.122936639 podStartE2EDuration="3.735444162s" podCreationTimestamp="2026-01-21 11:21:02 +0000 UTC" firstStartedPulling="2026-01-21 11:21:03.050562383 +0000 UTC m=+605.343591675" lastFinishedPulling="2026-01-21 11:21:04.663069906 +0000 UTC m=+606.956099198" observedRunningTime="2026-01-21 11:21:04.990183871 +0000 UTC m=+607.283213162" watchObservedRunningTime="2026-01-21 11:21:05.735444162 +0000 UTC m=+608.028473454" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.735697 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-w2j6n"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.736445 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.737734 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nm9q7" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.746998 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.747655 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.748904 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.758827 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-w2j6n"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.760702 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.763429 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-qjpv9"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.764089 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.768175 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h849b\" (UniqueName: \"kubernetes.io/projected/edf54909-b53e-44de-90bc-355e23bbc767-kube-api-access-h849b\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.768250 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.835672 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.841834 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.843414 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.843630 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-7tptd" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.844708 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.845604 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-dbus-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znknp\" (UniqueName: \"kubernetes.io/projected/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-kube-api-access-znknp\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870384 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-nmstate-lock\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: E0121 11:21:05.870396 4824 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870427 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h849b\" (UniqueName: \"kubernetes.io/projected/edf54909-b53e-44de-90bc-355e23bbc767-kube-api-access-h849b\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: E0121 11:21:05.870445 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair podName:edf54909-b53e-44de-90bc-355e23bbc767 nodeName:}" failed. No retries permitted until 2026-01-21 11:21:06.370431815 +0000 UTC m=+608.663461107 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair") pod "nmstate-webhook-8474b5b9d8-h6kkh" (UID: "edf54909-b53e-44de-90bc-355e23bbc767") : secret "openshift-nmstate-webhook" not found Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8dmq\" (UniqueName: \"kubernetes.io/projected/30b70686-bf05-4798-b5dc-3c7dc9cee56a-kube-api-access-j8dmq\") pod \"nmstate-metrics-54757c584b-w2j6n\" (UID: \"30b70686-bf05-4798-b5dc-3c7dc9cee56a\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870553 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv4cc\" (UniqueName: \"kubernetes.io/projected/1575ab9d-016f-403a-990b-3b4c4d53f6f1-kube-api-access-nv4cc\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870573 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1575ab9d-016f-403a-990b-3b4c4d53f6f1-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870632 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-ovs-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.893993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h849b\" (UniqueName: \"kubernetes.io/projected/edf54909-b53e-44de-90bc-355e23bbc767-kube-api-access-h849b\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-ovs-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-dbus-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znknp\" (UniqueName: \"kubernetes.io/projected/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-kube-api-access-znknp\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971448 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-nmstate-lock\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8dmq\" (UniqueName: \"kubernetes.io/projected/30b70686-bf05-4798-b5dc-3c7dc9cee56a-kube-api-access-j8dmq\") pod \"nmstate-metrics-54757c584b-w2j6n\" (UID: \"30b70686-bf05-4798-b5dc-3c7dc9cee56a\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv4cc\" (UniqueName: \"kubernetes.io/projected/1575ab9d-016f-403a-990b-3b4c4d53f6f1-kube-api-access-nv4cc\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1575ab9d-016f-403a-990b-3b4c4d53f6f1-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971653 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-dbus-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971211 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-ovs-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: E0121 11:21:05.971806 4824 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Jan 21 11:21:05 crc kubenswrapper[4824]: E0121 11:21:05.971849 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert podName:1575ab9d-016f-403a-990b-3b4c4d53f6f1 nodeName:}" failed. No retries permitted until 2026-01-21 11:21:06.471836561 +0000 UTC m=+608.764865853 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert") pod "nmstate-console-plugin-7754f76f8b-xjqs5" (UID: "1575ab9d-016f-403a-990b-3b4c4d53f6f1") : secret "plugin-serving-cert" not found Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971916 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-nmstate-lock\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.972287 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1575ab9d-016f-403a-990b-3b4c4d53f6f1-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.994640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv4cc\" (UniqueName: \"kubernetes.io/projected/1575ab9d-016f-403a-990b-3b4c4d53f6f1-kube-api-access-nv4cc\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.995126 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znknp\" (UniqueName: \"kubernetes.io/projected/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-kube-api-access-znknp\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.003540 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8dmq\" (UniqueName: \"kubernetes.io/projected/30b70686-bf05-4798-b5dc-3c7dc9cee56a-kube-api-access-j8dmq\") pod \"nmstate-metrics-54757c584b-w2j6n\" (UID: \"30b70686-bf05-4798-b5dc-3c7dc9cee56a\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.048205 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.076061 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.090179 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-68f8959f95-5v54n"] Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.090811 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: W0121 11:21:06.094314 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda01ecff0_9051_4ddb_8c8e_a4f379b6c823.slice/crio-28b033c87d7f0443ef7f4ad888ac03af97dfddab52c519537fc47d807edac1d3 WatchSource:0}: Error finding container 28b033c87d7f0443ef7f4ad888ac03af97dfddab52c519537fc47d807edac1d3: Status 404 returned error can't find the container with id 28b033c87d7f0443ef7f4ad888ac03af97dfddab52c519537fc47d807edac1d3 Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.105380 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-68f8959f95-5v54n"] Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-oauth-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-trusted-ca-bundle\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173767 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-oauth-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-service-ca\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173952 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.174034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48fpp\" (UniqueName: \"kubernetes.io/projected/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-kube-api-access-48fpp\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.228378 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-w2j6n"] Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274559 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48fpp\" (UniqueName: \"kubernetes.io/projected/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-kube-api-access-48fpp\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274683 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-oauth-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274699 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-trusted-ca-bundle\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274713 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-oauth-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-service-ca\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.275632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-service-ca\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.275658 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-oauth-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.275794 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-trusted-ca-bundle\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.275918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.277795 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-oauth-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.277984 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.287780 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48fpp\" (UniqueName: \"kubernetes.io/projected/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-kube-api-access-48fpp\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.375319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.378589 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.409541 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.476187 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.478770 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.661876 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.755187 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.761482 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-68f8959f95-5v54n"] Jan 21 11:21:06 crc kubenswrapper[4824]: W0121 11:21:06.766385 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bf3d16e_5b03_46d9_a09c_076fbea86f1d.slice/crio-f2bd8103f9f0316e4fa3b3650c5acf4310b28769b7af0d9f813d086f15b076e6 WatchSource:0}: Error finding container f2bd8103f9f0316e4fa3b3650c5acf4310b28769b7af0d9f813d086f15b076e6: Status 404 returned error can't find the container with id f2bd8103f9f0316e4fa3b3650c5acf4310b28769b7af0d9f813d086f15b076e6 Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.903282 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5"] Jan 21 11:21:06 crc kubenswrapper[4824]: W0121 11:21:06.912518 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1575ab9d_016f_403a_990b_3b4c4d53f6f1.slice/crio-4ff43fd17a851104dc40b9a7cffa012b1cff642cd009615e3cb6e93c57b2498f WatchSource:0}: Error finding container 4ff43fd17a851104dc40b9a7cffa012b1cff642cd009615e3cb6e93c57b2498f: Status 404 returned error can't find the container with id 4ff43fd17a851104dc40b9a7cffa012b1cff642cd009615e3cb6e93c57b2498f Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.990758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-68f8959f95-5v54n" event={"ID":"7bf3d16e-5b03-46d9-a09c-076fbea86f1d","Type":"ContainerStarted","Data":"f9d67e2a9ab1d7f288440eb669207a4003bbd4b49cdb9299daaa81316fdddd61"} Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.991117 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-68f8959f95-5v54n" event={"ID":"7bf3d16e-5b03-46d9-a09c-076fbea86f1d","Type":"ContainerStarted","Data":"f2bd8103f9f0316e4fa3b3650c5acf4310b28769b7af0d9f813d086f15b076e6"} Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.992456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" event={"ID":"1575ab9d-016f-403a-990b-3b4c4d53f6f1","Type":"ContainerStarted","Data":"4ff43fd17a851104dc40b9a7cffa012b1cff642cd009615e3cb6e93c57b2498f"} Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.994454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-qjpv9" event={"ID":"a01ecff0-9051-4ddb-8c8e-a4f379b6c823","Type":"ContainerStarted","Data":"28b033c87d7f0443ef7f4ad888ac03af97dfddab52c519537fc47d807edac1d3"} Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.995999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" event={"ID":"30b70686-bf05-4798-b5dc-3c7dc9cee56a","Type":"ContainerStarted","Data":"77bcedfac8711709d0029f27edf7cb259166fb4d4d40ca365fd9d40e7a422c44"} Jan 21 11:21:07 crc kubenswrapper[4824]: I0121 11:21:07.012066 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-68f8959f95-5v54n" podStartSLOduration=1.012028192 podStartE2EDuration="1.012028192s" podCreationTimestamp="2026-01-21 11:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:21:07.003775946 +0000 UTC m=+609.296805238" watchObservedRunningTime="2026-01-21 11:21:07.012028192 +0000 UTC m=+609.305057484" Jan 21 11:21:07 crc kubenswrapper[4824]: I0121 11:21:07.052263 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh"] Jan 21 11:21:08 crc kubenswrapper[4824]: I0121 11:21:08.003227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" event={"ID":"edf54909-b53e-44de-90bc-355e23bbc767","Type":"ContainerStarted","Data":"af018cf706cab903b3bc5faab1eb0afff311f9420b7439621c8d45b257a9ff07"} Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.008241 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" event={"ID":"30b70686-bf05-4798-b5dc-3c7dc9cee56a","Type":"ContainerStarted","Data":"b8ea1c7781977f653e93553dad86c7e755fa043f3b3876c6acb9122e2e388446"} Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.010583 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" event={"ID":"edf54909-b53e-44de-90bc-355e23bbc767","Type":"ContainerStarted","Data":"593c0deeb8c98823905656739795f6161a9cdf083e9c64bceec49e676cedb143"} Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.010704 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.011991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" event={"ID":"1575ab9d-016f-403a-990b-3b4c4d53f6f1","Type":"ContainerStarted","Data":"1b059417217898b40ec29c93b3b17a039f8ababce0e90ca359107aacd389cd30"} Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.013088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-qjpv9" event={"ID":"a01ecff0-9051-4ddb-8c8e-a4f379b6c823","Type":"ContainerStarted","Data":"24e5f66c433e84206edd442af9dec71753b7880999d768b147b02732caf21946"} Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.013226 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.026370 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" podStartSLOduration=3.034105345 podStartE2EDuration="4.026357168s" podCreationTimestamp="2026-01-21 11:21:05 +0000 UTC" firstStartedPulling="2026-01-21 11:21:07.052261514 +0000 UTC m=+609.345290806" lastFinishedPulling="2026-01-21 11:21:08.044513337 +0000 UTC m=+610.337542629" observedRunningTime="2026-01-21 11:21:09.024772539 +0000 UTC m=+611.317801831" watchObservedRunningTime="2026-01-21 11:21:09.026357168 +0000 UTC m=+611.319386460" Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.036716 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" podStartSLOduration=2.192533945 podStartE2EDuration="4.0367012s" podCreationTimestamp="2026-01-21 11:21:05 +0000 UTC" firstStartedPulling="2026-01-21 11:21:06.915380479 +0000 UTC m=+609.208409771" lastFinishedPulling="2026-01-21 11:21:08.759547733 +0000 UTC m=+611.052577026" observedRunningTime="2026-01-21 11:21:09.03436363 +0000 UTC m=+611.327392922" watchObservedRunningTime="2026-01-21 11:21:09.0367012 +0000 UTC m=+611.329730492" Jan 21 11:21:11 crc kubenswrapper[4824]: I0121 11:21:11.023072 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" event={"ID":"30b70686-bf05-4798-b5dc-3c7dc9cee56a","Type":"ContainerStarted","Data":"3d7e837c034443be3ebd5618c9a2e57a07267b1da725bf4eadf39c08a24b5208"} Jan 21 11:21:11 crc kubenswrapper[4824]: I0121 11:21:11.036413 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-qjpv9" podStartSLOduration=4.083457826 podStartE2EDuration="6.036401262s" podCreationTimestamp="2026-01-21 11:21:05 +0000 UTC" firstStartedPulling="2026-01-21 11:21:06.096059158 +0000 UTC m=+608.389088451" lastFinishedPulling="2026-01-21 11:21:08.049002594 +0000 UTC m=+610.342031887" observedRunningTime="2026-01-21 11:21:09.045900172 +0000 UTC m=+611.338929464" watchObservedRunningTime="2026-01-21 11:21:11.036401262 +0000 UTC m=+613.329430553" Jan 21 11:21:11 crc kubenswrapper[4824]: I0121 11:21:11.037644 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" podStartSLOduration=2.255024729 podStartE2EDuration="6.037636171s" podCreationTimestamp="2026-01-21 11:21:05 +0000 UTC" firstStartedPulling="2026-01-21 11:21:06.238621921 +0000 UTC m=+608.531651212" lastFinishedPulling="2026-01-21 11:21:10.021233362 +0000 UTC m=+612.314262654" observedRunningTime="2026-01-21 11:21:11.034036752 +0000 UTC m=+613.327066044" watchObservedRunningTime="2026-01-21 11:21:11.037636171 +0000 UTC m=+613.330665464" Jan 21 11:21:16 crc kubenswrapper[4824]: I0121 11:21:16.095474 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:16 crc kubenswrapper[4824]: I0121 11:21:16.410631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:16 crc kubenswrapper[4824]: I0121 11:21:16.410678 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:16 crc kubenswrapper[4824]: I0121 11:21:16.414480 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:17 crc kubenswrapper[4824]: I0121 11:21:17.051904 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:17 crc kubenswrapper[4824]: I0121 11:21:17.082349 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:21:26 crc kubenswrapper[4824]: I0121 11:21:26.666478 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.024296 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq"] Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.025543 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.026727 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.036063 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq"] Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.180784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmm2f\" (UniqueName: \"kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.180818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.180878 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.282101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmm2f\" (UniqueName: \"kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.282132 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.282184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.282603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.282643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.296244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmm2f\" (UniqueName: \"kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.337730 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.673022 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq"] Jan 21 11:21:36 crc kubenswrapper[4824]: I0121 11:21:36.122277 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9005d95-8246-435d-aaea-5c60d58175c6" containerID="27732c80d1e351e888b3afdbfb3574b4168cb28379f32229565b988ce91bccf7" exitCode=0 Jan 21 11:21:36 crc kubenswrapper[4824]: I0121 11:21:36.122351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" event={"ID":"d9005d95-8246-435d-aaea-5c60d58175c6","Type":"ContainerDied","Data":"27732c80d1e351e888b3afdbfb3574b4168cb28379f32229565b988ce91bccf7"} Jan 21 11:21:36 crc kubenswrapper[4824]: I0121 11:21:36.122477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" event={"ID":"d9005d95-8246-435d-aaea-5c60d58175c6","Type":"ContainerStarted","Data":"3169369e4043b11755f263f8f5ffe3764db94ba00d13ead4ec3ef289b12bce38"} Jan 21 11:21:38 crc kubenswrapper[4824]: I0121 11:21:38.131061 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9005d95-8246-435d-aaea-5c60d58175c6" containerID="67331e20e703337d471521971fe82b6954704b075e1322eff28c99c0a434fb47" exitCode=0 Jan 21 11:21:38 crc kubenswrapper[4824]: I0121 11:21:38.131091 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" event={"ID":"d9005d95-8246-435d-aaea-5c60d58175c6","Type":"ContainerDied","Data":"67331e20e703337d471521971fe82b6954704b075e1322eff28c99c0a434fb47"} Jan 21 11:21:39 crc kubenswrapper[4824]: I0121 11:21:39.135920 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9005d95-8246-435d-aaea-5c60d58175c6" containerID="bfd4bec6d81589bdf71cf6fa238cbe1dc1a1166fc05f16b127d0eb3075647620" exitCode=0 Jan 21 11:21:39 crc kubenswrapper[4824]: I0121 11:21:39.135984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" event={"ID":"d9005d95-8246-435d-aaea-5c60d58175c6","Type":"ContainerDied","Data":"bfd4bec6d81589bdf71cf6fa238cbe1dc1a1166fc05f16b127d0eb3075647620"} Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.305327 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.431625 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle\") pod \"d9005d95-8246-435d-aaea-5c60d58175c6\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.431667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmm2f\" (UniqueName: \"kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f\") pod \"d9005d95-8246-435d-aaea-5c60d58175c6\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.431705 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util\") pod \"d9005d95-8246-435d-aaea-5c60d58175c6\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.432502 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle" (OuterVolumeSpecName: "bundle") pod "d9005d95-8246-435d-aaea-5c60d58175c6" (UID: "d9005d95-8246-435d-aaea-5c60d58175c6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.435948 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f" (OuterVolumeSpecName: "kube-api-access-tmm2f") pod "d9005d95-8246-435d-aaea-5c60d58175c6" (UID: "d9005d95-8246-435d-aaea-5c60d58175c6"). InnerVolumeSpecName "kube-api-access-tmm2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.441590 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util" (OuterVolumeSpecName: "util") pod "d9005d95-8246-435d-aaea-5c60d58175c6" (UID: "d9005d95-8246-435d-aaea-5c60d58175c6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.532694 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.532719 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmm2f\" (UniqueName: \"kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.532729 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:41 crc kubenswrapper[4824]: I0121 11:21:41.145123 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" event={"ID":"d9005d95-8246-435d-aaea-5c60d58175c6","Type":"ContainerDied","Data":"3169369e4043b11755f263f8f5ffe3764db94ba00d13ead4ec3ef289b12bce38"} Jan 21 11:21:41 crc kubenswrapper[4824]: I0121 11:21:41.145155 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3169369e4043b11755f263f8f5ffe3764db94ba00d13ead4ec3ef289b12bce38" Jan 21 11:21:41 crc kubenswrapper[4824]: I0121 11:21:41.145171 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.106596 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-nwlwd" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" containerID="cri-o://3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7" gracePeriod=15 Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.384074 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-nwlwd_c585ef2f-3e29-4b81-918f-11b075006ca0/console/0.log" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.384137 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552190 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8qjw\" (UniqueName: \"kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552276 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552303 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552338 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552385 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552402 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552871 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca" (OuterVolumeSpecName: "service-ca") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.553148 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config" (OuterVolumeSpecName: "console-config") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.553310 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.553690 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.554080 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.554147 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.554161 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.554171 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.556849 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.556928 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw" (OuterVolumeSpecName: "kube-api-access-q8qjw") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "kube-api-access-q8qjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.557022 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.654804 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.655319 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8qjw\" (UniqueName: \"kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.655388 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154739 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-nwlwd_c585ef2f-3e29-4b81-918f-11b075006ca0/console/0.log" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154790 4824 generic.go:334] "Generic (PLEG): container finished" podID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerID="3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7" exitCode=2 Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154819 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nwlwd" event={"ID":"c585ef2f-3e29-4b81-918f-11b075006ca0","Type":"ContainerDied","Data":"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7"} Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154844 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nwlwd" event={"ID":"c585ef2f-3e29-4b81-918f-11b075006ca0","Type":"ContainerDied","Data":"0b16adc0798e89f22e1034c180951bc349e7c3b376c2bea592b22d2474b2bc5e"} Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154859 4824 scope.go:117] "RemoveContainer" containerID="3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154865 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.170339 4824 scope.go:117] "RemoveContainer" containerID="3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7" Jan 21 11:21:43 crc kubenswrapper[4824]: E0121 11:21:43.171519 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7\": container with ID starting with 3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7 not found: ID does not exist" containerID="3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.171572 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7"} err="failed to get container status \"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7\": rpc error: code = NotFound desc = could not find container \"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7\": container with ID starting with 3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7 not found: ID does not exist" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.181752 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.184411 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:21:44 crc kubenswrapper[4824]: I0121 11:21:44.053299 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" path="/var/lib/kubelet/pods/c585ef2f-3e29-4b81-918f-11b075006ca0/volumes" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.470615 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct"] Jan 21 11:21:49 crc kubenswrapper[4824]: E0121 11:21:49.471142 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="util" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471155 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="util" Jan 21 11:21:49 crc kubenswrapper[4824]: E0121 11:21:49.471163 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471168 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" Jan 21 11:21:49 crc kubenswrapper[4824]: E0121 11:21:49.471189 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="extract" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471194 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="extract" Jan 21 11:21:49 crc kubenswrapper[4824]: E0121 11:21:49.471200 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="pull" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471206 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="pull" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471315 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="extract" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471324 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471666 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.474827 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.475013 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.475180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-485hc" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.475026 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.475057 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.480554 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct"] Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.527061 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-webhook-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.527117 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5lrt\" (UniqueName: \"kubernetes.io/projected/714d2131-afc8-4828-840f-e2f0ce727d5d-kube-api-access-v5lrt\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.527147 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-apiservice-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.627843 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-webhook-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.627884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5lrt\" (UniqueName: \"kubernetes.io/projected/714d2131-afc8-4828-840f-e2f0ce727d5d-kube-api-access-v5lrt\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.627908 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-apiservice-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.633183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-apiservice-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.637521 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-webhook-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.641873 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5lrt\" (UniqueName: \"kubernetes.io/projected/714d2131-afc8-4828-840f-e2f0ce727d5d-kube-api-access-v5lrt\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.692250 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5"] Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.692864 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.694293 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.694315 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.694508 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-j4wrz" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.704214 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5"] Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.729260 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-apiservice-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.729314 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-webhook-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.729455 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzb82\" (UniqueName: \"kubernetes.io/projected/81fed5ab-ec3d-46f0-8998-199393bffba0-kube-api-access-wzb82\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.785946 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.830398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-apiservice-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.830648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-webhook-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.830692 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzb82\" (UniqueName: \"kubernetes.io/projected/81fed5ab-ec3d-46f0-8998-199393bffba0-kube-api-access-wzb82\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.837560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-webhook-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.847783 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-apiservice-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.848090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzb82\" (UniqueName: \"kubernetes.io/projected/81fed5ab-ec3d-46f0-8998-199393bffba0-kube-api-access-wzb82\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:50 crc kubenswrapper[4824]: I0121 11:21:50.010168 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:50 crc kubenswrapper[4824]: I0121 11:21:50.158721 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5"] Jan 21 11:21:50 crc kubenswrapper[4824]: W0121 11:21:50.163688 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81fed5ab_ec3d_46f0_8998_199393bffba0.slice/crio-e13fd4ef1e0aa49f805e3011bcab6e7503dfccd9f20ebd3374b90352ef667dd8 WatchSource:0}: Error finding container e13fd4ef1e0aa49f805e3011bcab6e7503dfccd9f20ebd3374b90352ef667dd8: Status 404 returned error can't find the container with id e13fd4ef1e0aa49f805e3011bcab6e7503dfccd9f20ebd3374b90352ef667dd8 Jan 21 11:21:50 crc kubenswrapper[4824]: I0121 11:21:50.176708 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct"] Jan 21 11:21:50 crc kubenswrapper[4824]: W0121 11:21:50.184179 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod714d2131_afc8_4828_840f_e2f0ce727d5d.slice/crio-02da55d074eb6b9f3efed95f9be4a2b12e2ee23fabe98f8b1db6fddea084e66f WatchSource:0}: Error finding container 02da55d074eb6b9f3efed95f9be4a2b12e2ee23fabe98f8b1db6fddea084e66f: Status 404 returned error can't find the container with id 02da55d074eb6b9f3efed95f9be4a2b12e2ee23fabe98f8b1db6fddea084e66f Jan 21 11:21:50 crc kubenswrapper[4824]: I0121 11:21:50.191892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" event={"ID":"81fed5ab-ec3d-46f0-8998-199393bffba0","Type":"ContainerStarted","Data":"e13fd4ef1e0aa49f805e3011bcab6e7503dfccd9f20ebd3374b90352ef667dd8"} Jan 21 11:21:51 crc kubenswrapper[4824]: I0121 11:21:51.197458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" event={"ID":"714d2131-afc8-4828-840f-e2f0ce727d5d","Type":"ContainerStarted","Data":"02da55d074eb6b9f3efed95f9be4a2b12e2ee23fabe98f8b1db6fddea084e66f"} Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.215051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" event={"ID":"714d2131-afc8-4828-840f-e2f0ce727d5d","Type":"ContainerStarted","Data":"946b563520e00ffbab367928e26c2cfb8af9b079810c7e19ceaa7ee88b6e11e5"} Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.215391 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.217172 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" event={"ID":"81fed5ab-ec3d-46f0-8998-199393bffba0","Type":"ContainerStarted","Data":"9f9097f18e9d26887efe54591063136fb92d1431b65090d400d4bb491708021c"} Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.217322 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.229017 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" podStartSLOduration=2.20324148 podStartE2EDuration="5.229005823s" podCreationTimestamp="2026-01-21 11:21:49 +0000 UTC" firstStartedPulling="2026-01-21 11:21:50.193665085 +0000 UTC m=+652.486694368" lastFinishedPulling="2026-01-21 11:21:53.219429418 +0000 UTC m=+655.512458711" observedRunningTime="2026-01-21 11:21:54.227360286 +0000 UTC m=+656.520389578" watchObservedRunningTime="2026-01-21 11:21:54.229005823 +0000 UTC m=+656.522035115" Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.245212 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" podStartSLOduration=2.184176941 podStartE2EDuration="5.245196505s" podCreationTimestamp="2026-01-21 11:21:49 +0000 UTC" firstStartedPulling="2026-01-21 11:21:50.164992755 +0000 UTC m=+652.458022047" lastFinishedPulling="2026-01-21 11:21:53.226012319 +0000 UTC m=+655.519041611" observedRunningTime="2026-01-21 11:21:54.242161865 +0000 UTC m=+656.535191157" watchObservedRunningTime="2026-01-21 11:21:54.245196505 +0000 UTC m=+656.538225797" Jan 21 11:22:10 crc kubenswrapper[4824]: I0121 11:22:10.013690 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:22:29 crc kubenswrapper[4824]: I0121 11:22:29.788874 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.341128 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-9q9wj"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.342907 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.344481 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-dgcz7" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.345044 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.345476 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.349122 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.350065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.351671 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.354840 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.374979 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.375100 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jxvb\" (UniqueName: \"kubernetes.io/projected/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-kube-api-access-5jxvb\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.417791 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-8tsb9"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.418990 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.422807 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-bbv7p"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.423281 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.423443 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.423357 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-2cwg7" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.423758 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.423876 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.426728 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.430889 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-bbv7p"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476697 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-metrics-certs\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476739 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6bd9\" (UniqueName: \"kubernetes.io/projected/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-kube-api-access-n6bd9\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476760 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metallb-excludel2\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476796 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-reloader\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476810 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-cert\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476831 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jxvb\" (UniqueName: \"kubernetes.io/projected/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-kube-api-access-5jxvb\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476855 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8zz4\" (UniqueName: \"kubernetes.io/projected/d30695b8-6f6a-48f0-88ed-9181484634b8-kube-api-access-s8zz4\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476982 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-startup\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477037 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf6fn\" (UniqueName: \"kubernetes.io/projected/e38576f0-c423-497a-a30d-60c2e8edd93a-kube-api-access-xf6fn\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477088 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-conf\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477105 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477128 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-sockets\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.492709 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.492814 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jxvb\" (UniqueName: \"kubernetes.io/projected/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-kube-api-access-5jxvb\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578290 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578358 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf6fn\" (UniqueName: \"kubernetes.io/projected/e38576f0-c423-497a-a30d-60c2e8edd93a-kube-api-access-xf6fn\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-conf\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578404 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-sockets\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578441 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-metrics-certs\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6bd9\" (UniqueName: \"kubernetes.io/projected/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-kube-api-access-n6bd9\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metallb-excludel2\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578488 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578506 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-reloader\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-cert\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578542 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8zz4\" (UniqueName: \"kubernetes.io/projected/d30695b8-6f6a-48f0-88ed-9181484634b8-kube-api-access-s8zz4\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578600 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-startup\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.579420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-startup\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.579641 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.579726 4824 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.579769 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs podName:baf5db51-d3ac-4b97-9c5f-afc598d59ab7 nodeName:}" failed. No retries permitted until 2026-01-21 11:22:31.079754946 +0000 UTC m=+693.372784238 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs") pod "speaker-8tsb9" (UID: "baf5db51-d3ac-4b97-9c5f-afc598d59ab7") : secret "speaker-certs-secret" not found Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.580283 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-conf\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.580451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-sockets\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.580984 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-reloader\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.580983 4824 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.581097 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs podName:e38576f0-c423-497a-a30d-60c2e8edd93a nodeName:}" failed. No retries permitted until 2026-01-21 11:22:31.081086511 +0000 UTC m=+693.374115802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs") pod "frr-k8s-9q9wj" (UID: "e38576f0-c423-497a-a30d-60c2e8edd93a") : secret "frr-k8s-certs-secret" not found Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.581049 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.581128 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist podName:baf5db51-d3ac-4b97-9c5f-afc598d59ab7 nodeName:}" failed. No retries permitted until 2026-01-21 11:22:31.081122678 +0000 UTC m=+693.374151970 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist") pod "speaker-8tsb9" (UID: "baf5db51-d3ac-4b97-9c5f-afc598d59ab7") : secret "metallb-memberlist" not found Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.581647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metallb-excludel2\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.582539 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.583853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-metrics-certs\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.593273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf6fn\" (UniqueName: \"kubernetes.io/projected/e38576f0-c423-497a-a30d-60c2e8edd93a-kube-api-access-xf6fn\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.593777 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-cert\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.595358 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8zz4\" (UniqueName: \"kubernetes.io/projected/d30695b8-6f6a-48f0-88ed-9181484634b8-kube-api-access-s8zz4\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.596274 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6bd9\" (UniqueName: \"kubernetes.io/projected/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-kube-api-access-n6bd9\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.678821 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.741016 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.083694 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:31 crc kubenswrapper[4824]: E0121 11:22:31.083776 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.083989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:31 crc kubenswrapper[4824]: E0121 11:22:31.084034 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist podName:baf5db51-d3ac-4b97-9c5f-afc598d59ab7 nodeName:}" failed. No retries permitted until 2026-01-21 11:22:32.084015757 +0000 UTC m=+694.377045049 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist") pod "speaker-8tsb9" (UID: "baf5db51-d3ac-4b97-9c5f-afc598d59ab7") : secret "metallb-memberlist" not found Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.084054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.088394 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb"] Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.093186 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.093218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.152264 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-bbv7p"] Jan 21 11:22:31 crc kubenswrapper[4824]: W0121 11:22:31.158567 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd30695b8_6f6a_48f0_88ed_9181484634b8.slice/crio-e91b7eab093938e8669c4c30f13b245b7708535eb6b924ed385a4505933e198e WatchSource:0}: Error finding container e91b7eab093938e8669c4c30f13b245b7708535eb6b924ed385a4505933e198e: Status 404 returned error can't find the container with id e91b7eab093938e8669c4c30f13b245b7708535eb6b924ed385a4505933e198e Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.258224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.381999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-bbv7p" event={"ID":"d30695b8-6f6a-48f0-88ed-9181484634b8","Type":"ContainerStarted","Data":"6941c121390bf3d300db662a46888f442a975e5dcb6064e1bbafeffed095a417"} Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.382040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-bbv7p" event={"ID":"d30695b8-6f6a-48f0-88ed-9181484634b8","Type":"ContainerStarted","Data":"cf5b3ae27e69bc46a35b70866df64d613f991437dc86f83bc3a382de36b29dd6"} Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.382053 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-bbv7p" event={"ID":"d30695b8-6f6a-48f0-88ed-9181484634b8","Type":"ContainerStarted","Data":"e91b7eab093938e8669c4c30f13b245b7708535eb6b924ed385a4505933e198e"} Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.382133 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.383072 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"12be331cc3262346a8fa63374a3426fa6b02cfc67db2a8b9421fb800f4d5536f"} Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.383936 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" event={"ID":"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a","Type":"ContainerStarted","Data":"e421d48f3b6d0a95ac7a49d2adc2bf51886d1df12f496efea31ce2742524a525"} Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.395342 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-bbv7p" podStartSLOduration=1.395330277 podStartE2EDuration="1.395330277s" podCreationTimestamp="2026-01-21 11:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:22:31.392841518 +0000 UTC m=+693.685870810" watchObservedRunningTime="2026-01-21 11:22:31.395330277 +0000 UTC m=+693.688359569" Jan 21 11:22:32 crc kubenswrapper[4824]: I0121 11:22:32.093758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:32 crc kubenswrapper[4824]: I0121 11:22:32.110842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:32 crc kubenswrapper[4824]: I0121 11:22:32.235201 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8tsb9" Jan 21 11:22:32 crc kubenswrapper[4824]: I0121 11:22:32.401764 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8tsb9" event={"ID":"baf5db51-d3ac-4b97-9c5f-afc598d59ab7","Type":"ContainerStarted","Data":"00408b77679c420ad9a736839489c2fafe53e9a4cdd59abe8d5fd7728fd5286a"} Jan 21 11:22:33 crc kubenswrapper[4824]: I0121 11:22:33.411348 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8tsb9" event={"ID":"baf5db51-d3ac-4b97-9c5f-afc598d59ab7","Type":"ContainerStarted","Data":"d8500e599fa9c0d231db01e19baeca992b240d084b6c51b01ea472922e43afa2"} Jan 21 11:22:33 crc kubenswrapper[4824]: I0121 11:22:33.411673 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8tsb9" event={"ID":"baf5db51-d3ac-4b97-9c5f-afc598d59ab7","Type":"ContainerStarted","Data":"cbffadfc9555b5b289cbd0ae205d85f95a914b260f3dfee6102d45e8d3ee91c3"} Jan 21 11:22:33 crc kubenswrapper[4824]: I0121 11:22:33.411690 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-8tsb9" Jan 21 11:22:33 crc kubenswrapper[4824]: I0121 11:22:33.426988 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-8tsb9" podStartSLOduration=3.426972368 podStartE2EDuration="3.426972368s" podCreationTimestamp="2026-01-21 11:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:22:33.424863806 +0000 UTC m=+695.717893099" watchObservedRunningTime="2026-01-21 11:22:33.426972368 +0000 UTC m=+695.720001660" Jan 21 11:22:37 crc kubenswrapper[4824]: I0121 11:22:37.439842 4824 generic.go:334] "Generic (PLEG): container finished" podID="e38576f0-c423-497a-a30d-60c2e8edd93a" containerID="0a95b49dd0f30bdcb9480450ec3e995683dcc1c5bef8adea27f3a85a5d236c1d" exitCode=0 Jan 21 11:22:37 crc kubenswrapper[4824]: I0121 11:22:37.439987 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerDied","Data":"0a95b49dd0f30bdcb9480450ec3e995683dcc1c5bef8adea27f3a85a5d236c1d"} Jan 21 11:22:37 crc kubenswrapper[4824]: I0121 11:22:37.443813 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" event={"ID":"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a","Type":"ContainerStarted","Data":"344ca599880fd08d1fb66b1e7118337f7fdf955b7f9b5d0d142f0d2aee4d5487"} Jan 21 11:22:37 crc kubenswrapper[4824]: I0121 11:22:37.443988 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:37 crc kubenswrapper[4824]: I0121 11:22:37.470942 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" podStartSLOduration=1.917495246 podStartE2EDuration="7.470924236s" podCreationTimestamp="2026-01-21 11:22:30 +0000 UTC" firstStartedPulling="2026-01-21 11:22:31.095394333 +0000 UTC m=+693.388423614" lastFinishedPulling="2026-01-21 11:22:36.648823312 +0000 UTC m=+698.941852604" observedRunningTime="2026-01-21 11:22:37.470544929 +0000 UTC m=+699.763574221" watchObservedRunningTime="2026-01-21 11:22:37.470924236 +0000 UTC m=+699.763953528" Jan 21 11:22:38 crc kubenswrapper[4824]: I0121 11:22:38.449558 4824 generic.go:334] "Generic (PLEG): container finished" podID="e38576f0-c423-497a-a30d-60c2e8edd93a" containerID="02f6cd335f7b59d9369a0b173b948cb2181ed1b07a21195f07787f05898d6c8f" exitCode=0 Jan 21 11:22:38 crc kubenswrapper[4824]: I0121 11:22:38.449602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerDied","Data":"02f6cd335f7b59d9369a0b173b948cb2181ed1b07a21195f07787f05898d6c8f"} Jan 21 11:22:39 crc kubenswrapper[4824]: I0121 11:22:39.456453 4824 generic.go:334] "Generic (PLEG): container finished" podID="e38576f0-c423-497a-a30d-60c2e8edd93a" containerID="8e5df092e2d0209c1dd968162c6bb2a85f1a4e9f92796f3723722c15b87ebeb3" exitCode=0 Jan 21 11:22:39 crc kubenswrapper[4824]: I0121 11:22:39.456498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerDied","Data":"8e5df092e2d0209c1dd968162c6bb2a85f1a4e9f92796f3723722c15b87ebeb3"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"dceffcf088cbc712e691f9a86c11e2343a15757b03b28b0eecd852c705dd889c"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"61a0f960fe6522f77e86258dc114d4a59ad0c0f8644703aea226afefae7a28df"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465484 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"9ba551b63fd42c126b2373cfc52ce8a5958522b4cd4bd230ea86d5c8ac0dcfc0"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465498 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465507 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"276db60e826eda049265f4eeb3b091ef6c8beb0d2e747421f737f99e465f7838"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465515 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"077e5c87fa526fc502e122e29f0cb39c220b47fe3bb8ad2146bd1f65ac9b5519"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"eac894334aecaf0c6194362c91ce71a737232cd8eacc273c58adeca41bc80a23"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.481827 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-9q9wj" podStartSLOduration=5.160693349 podStartE2EDuration="10.481815485s" podCreationTimestamp="2026-01-21 11:22:30 +0000 UTC" firstStartedPulling="2026-01-21 11:22:31.333815642 +0000 UTC m=+693.626844924" lastFinishedPulling="2026-01-21 11:22:36.654937767 +0000 UTC m=+698.947967060" observedRunningTime="2026-01-21 11:22:40.481201355 +0000 UTC m=+702.774230647" watchObservedRunningTime="2026-01-21 11:22:40.481815485 +0000 UTC m=+702.774844777" Jan 21 11:22:41 crc kubenswrapper[4824]: I0121 11:22:41.259054 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:41 crc kubenswrapper[4824]: I0121 11:22:41.288052 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:42 crc kubenswrapper[4824]: I0121 11:22:42.238549 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-8tsb9" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.326361 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.327008 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.329120 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-dlsfl" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.330282 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.333767 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.334574 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.460138 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6bsj\" (UniqueName: \"kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj\") pod \"openstack-operator-index-2xq8c\" (UID: \"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d\") " pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.561494 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6bsj\" (UniqueName: \"kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj\") pod \"openstack-operator-index-2xq8c\" (UID: \"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d\") " pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.577836 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6bsj\" (UniqueName: \"kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj\") pod \"openstack-operator-index-2xq8c\" (UID: \"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d\") " pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.642528 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.978536 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:45 crc kubenswrapper[4824]: I0121 11:22:45.489273 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xq8c" event={"ID":"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d","Type":"ContainerStarted","Data":"65770a4cd5ba23f67f4e7e3da13403242a193d33807de294fba0702d80bb3eec"} Jan 21 11:22:46 crc kubenswrapper[4824]: I0121 11:22:46.065275 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:22:46 crc kubenswrapper[4824]: I0121 11:22:46.065316 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:22:47 crc kubenswrapper[4824]: I0121 11:22:47.499316 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xq8c" event={"ID":"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d","Type":"ContainerStarted","Data":"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266"} Jan 21 11:22:47 crc kubenswrapper[4824]: I0121 11:22:47.511918 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2xq8c" podStartSLOduration=1.48636971 podStartE2EDuration="3.511903518s" podCreationTimestamp="2026-01-21 11:22:44 +0000 UTC" firstStartedPulling="2026-01-21 11:22:44.986417396 +0000 UTC m=+707.279446687" lastFinishedPulling="2026-01-21 11:22:47.011951202 +0000 UTC m=+709.304980495" observedRunningTime="2026-01-21 11:22:47.509012018 +0000 UTC m=+709.802041310" watchObservedRunningTime="2026-01-21 11:22:47.511903518 +0000 UTC m=+709.804932810" Jan 21 11:22:47 crc kubenswrapper[4824]: I0121 11:22:47.709671 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.313047 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zggxm"] Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.313640 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.317857 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zggxm"] Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.407994 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfdz9\" (UniqueName: \"kubernetes.io/projected/78eb1906-d9e6-49f7-a512-ed3cc0a07da7-kube-api-access-dfdz9\") pod \"openstack-operator-index-zggxm\" (UID: \"78eb1906-d9e6-49f7-a512-ed3cc0a07da7\") " pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.509349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfdz9\" (UniqueName: \"kubernetes.io/projected/78eb1906-d9e6-49f7-a512-ed3cc0a07da7-kube-api-access-dfdz9\") pod \"openstack-operator-index-zggxm\" (UID: \"78eb1906-d9e6-49f7-a512-ed3cc0a07da7\") " pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.523699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfdz9\" (UniqueName: \"kubernetes.io/projected/78eb1906-d9e6-49f7-a512-ed3cc0a07da7-kube-api-access-dfdz9\") pod \"openstack-operator-index-zggxm\" (UID: \"78eb1906-d9e6-49f7-a512-ed3cc0a07da7\") " pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.627122 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.955828 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zggxm"] Jan 21 11:22:48 crc kubenswrapper[4824]: W0121 11:22:48.958414 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78eb1906_d9e6_49f7_a512_ed3cc0a07da7.slice/crio-817c80c60202a27b25381cfe512730c140beb8ad74df7f8cc7375b67d651a6ea WatchSource:0}: Error finding container 817c80c60202a27b25381cfe512730c140beb8ad74df7f8cc7375b67d651a6ea: Status 404 returned error can't find the container with id 817c80c60202a27b25381cfe512730c140beb8ad74df7f8cc7375b67d651a6ea Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.509165 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zggxm" event={"ID":"78eb1906-d9e6-49f7-a512-ed3cc0a07da7","Type":"ContainerStarted","Data":"88794dbd990089c8d2da8afa3daa81d9dd2222adf0cc7353a709e9d78811072d"} Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.509216 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zggxm" event={"ID":"78eb1906-d9e6-49f7-a512-ed3cc0a07da7","Type":"ContainerStarted","Data":"817c80c60202a27b25381cfe512730c140beb8ad74df7f8cc7375b67d651a6ea"} Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.509255 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-2xq8c" podUID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" containerName="registry-server" containerID="cri-o://127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266" gracePeriod=2 Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.525754 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zggxm" podStartSLOduration=1.142007744 podStartE2EDuration="1.525737543s" podCreationTimestamp="2026-01-21 11:22:48 +0000 UTC" firstStartedPulling="2026-01-21 11:22:48.961313879 +0000 UTC m=+711.254343171" lastFinishedPulling="2026-01-21 11:22:49.345043678 +0000 UTC m=+711.638072970" observedRunningTime="2026-01-21 11:22:49.522488348 +0000 UTC m=+711.815517640" watchObservedRunningTime="2026-01-21 11:22:49.525737543 +0000 UTC m=+711.818766835" Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.800908 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.925879 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6bsj\" (UniqueName: \"kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj\") pod \"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d\" (UID: \"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d\") " Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.930033 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj" (OuterVolumeSpecName: "kube-api-access-p6bsj") pod "3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" (UID: "3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d"). InnerVolumeSpecName "kube-api-access-p6bsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.026876 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6bsj\" (UniqueName: \"kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj\") on node \"crc\" DevicePath \"\"" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.514250 4824 generic.go:334] "Generic (PLEG): container finished" podID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" containerID="127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266" exitCode=0 Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.514286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xq8c" event={"ID":"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d","Type":"ContainerDied","Data":"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266"} Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.514318 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.514337 4824 scope.go:117] "RemoveContainer" containerID="127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.514325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xq8c" event={"ID":"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d","Type":"ContainerDied","Data":"65770a4cd5ba23f67f4e7e3da13403242a193d33807de294fba0702d80bb3eec"} Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.527942 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.529163 4824 scope.go:117] "RemoveContainer" containerID="127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266" Jan 21 11:22:50 crc kubenswrapper[4824]: E0121 11:22:50.529466 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266\": container with ID starting with 127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266 not found: ID does not exist" containerID="127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.529512 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266"} err="failed to get container status \"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266\": rpc error: code = NotFound desc = could not find container \"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266\": container with ID starting with 127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266 not found: ID does not exist" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.531178 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.683807 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.744380 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:51 crc kubenswrapper[4824]: I0121 11:22:51.260733 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:52 crc kubenswrapper[4824]: I0121 11:22:52.053953 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" path="/var/lib/kubelet/pods/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d/volumes" Jan 21 11:22:58 crc kubenswrapper[4824]: I0121 11:22:58.628202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:58 crc kubenswrapper[4824]: I0121 11:22:58.628387 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:58 crc kubenswrapper[4824]: I0121 11:22:58.647797 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:59 crc kubenswrapper[4824]: I0121 11:22:59.572301 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.436062 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr"] Jan 21 11:23:07 crc kubenswrapper[4824]: E0121 11:23:07.436602 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" containerName="registry-server" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.436614 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" containerName="registry-server" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.436745 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" containerName="registry-server" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.437436 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.439219 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-gw7z7" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.447703 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr"] Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.532449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.532554 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clhll\" (UniqueName: \"kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.532588 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.633464 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.633555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clhll\" (UniqueName: \"kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.633579 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.633938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.634050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.649042 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clhll\" (UniqueName: \"kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.750571 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:08 crc kubenswrapper[4824]: I0121 11:23:08.089073 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr"] Jan 21 11:23:08 crc kubenswrapper[4824]: W0121 11:23:08.093812 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72ab26cb_149e_4825_9587_3acfa66a368f.slice/crio-3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83 WatchSource:0}: Error finding container 3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83: Status 404 returned error can't find the container with id 3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83 Jan 21 11:23:08 crc kubenswrapper[4824]: I0121 11:23:08.594389 4824 generic.go:334] "Generic (PLEG): container finished" podID="72ab26cb-149e-4825-9587-3acfa66a368f" containerID="fc23b7440efb654b0a6f0f66596889abdd1f9ef8f8fc616eb9d053707e783e4b" exitCode=0 Jan 21 11:23:08 crc kubenswrapper[4824]: I0121 11:23:08.594436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" event={"ID":"72ab26cb-149e-4825-9587-3acfa66a368f","Type":"ContainerDied","Data":"fc23b7440efb654b0a6f0f66596889abdd1f9ef8f8fc616eb9d053707e783e4b"} Jan 21 11:23:08 crc kubenswrapper[4824]: I0121 11:23:08.594460 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" event={"ID":"72ab26cb-149e-4825-9587-3acfa66a368f","Type":"ContainerStarted","Data":"3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83"} Jan 21 11:23:09 crc kubenswrapper[4824]: I0121 11:23:09.599404 4824 generic.go:334] "Generic (PLEG): container finished" podID="72ab26cb-149e-4825-9587-3acfa66a368f" containerID="c380719aa42b9f83e4a40c333480e1664ba50c6d0ecce4d8e13a8c5281aac946" exitCode=0 Jan 21 11:23:09 crc kubenswrapper[4824]: I0121 11:23:09.599464 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" event={"ID":"72ab26cb-149e-4825-9587-3acfa66a368f","Type":"ContainerDied","Data":"c380719aa42b9f83e4a40c333480e1664ba50c6d0ecce4d8e13a8c5281aac946"} Jan 21 11:23:10 crc kubenswrapper[4824]: I0121 11:23:10.609859 4824 generic.go:334] "Generic (PLEG): container finished" podID="72ab26cb-149e-4825-9587-3acfa66a368f" containerID="7b18087c00cc8c90c5738a6f8ee462f79b3296105378b8976f0a6d0af38ecccf" exitCode=0 Jan 21 11:23:10 crc kubenswrapper[4824]: I0121 11:23:10.609901 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" event={"ID":"72ab26cb-149e-4825-9587-3acfa66a368f","Type":"ContainerDied","Data":"7b18087c00cc8c90c5738a6f8ee462f79b3296105378b8976f0a6d0af38ecccf"} Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.797531 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.978247 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle\") pod \"72ab26cb-149e-4825-9587-3acfa66a368f\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.978351 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clhll\" (UniqueName: \"kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll\") pod \"72ab26cb-149e-4825-9587-3acfa66a368f\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.978417 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util\") pod \"72ab26cb-149e-4825-9587-3acfa66a368f\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.978844 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle" (OuterVolumeSpecName: "bundle") pod "72ab26cb-149e-4825-9587-3acfa66a368f" (UID: "72ab26cb-149e-4825-9587-3acfa66a368f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.982938 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll" (OuterVolumeSpecName: "kube-api-access-clhll") pod "72ab26cb-149e-4825-9587-3acfa66a368f" (UID: "72ab26cb-149e-4825-9587-3acfa66a368f"). InnerVolumeSpecName "kube-api-access-clhll". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.988254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util" (OuterVolumeSpecName: "util") pod "72ab26cb-149e-4825-9587-3acfa66a368f" (UID: "72ab26cb-149e-4825-9587-3acfa66a368f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.080931 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clhll\" (UniqueName: \"kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll\") on node \"crc\" DevicePath \"\"" Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.081001 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util\") on node \"crc\" DevicePath \"\"" Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.081015 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.620230 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" event={"ID":"72ab26cb-149e-4825-9587-3acfa66a368f","Type":"ContainerDied","Data":"3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83"} Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.620275 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83" Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.620337 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.830760 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b"] Jan 21 11:23:14 crc kubenswrapper[4824]: E0121 11:23:14.831179 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="pull" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.831191 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="pull" Jan 21 11:23:14 crc kubenswrapper[4824]: E0121 11:23:14.831200 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="util" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.831205 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="util" Jan 21 11:23:14 crc kubenswrapper[4824]: E0121 11:23:14.831219 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="extract" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.831226 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="extract" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.831323 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="extract" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.831682 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.834137 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-9hmt7" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.856112 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b"] Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.916437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2hv9\" (UniqueName: \"kubernetes.io/projected/35334861-8e16-4b16-8cf5-167a049836af-kube-api-access-d2hv9\") pod \"openstack-operator-controller-init-dbdbd6855-gkm5b\" (UID: \"35334861-8e16-4b16-8cf5-167a049836af\") " pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:15 crc kubenswrapper[4824]: I0121 11:23:15.017433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2hv9\" (UniqueName: \"kubernetes.io/projected/35334861-8e16-4b16-8cf5-167a049836af-kube-api-access-d2hv9\") pod \"openstack-operator-controller-init-dbdbd6855-gkm5b\" (UID: \"35334861-8e16-4b16-8cf5-167a049836af\") " pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:15 crc kubenswrapper[4824]: I0121 11:23:15.035179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2hv9\" (UniqueName: \"kubernetes.io/projected/35334861-8e16-4b16-8cf5-167a049836af-kube-api-access-d2hv9\") pod \"openstack-operator-controller-init-dbdbd6855-gkm5b\" (UID: \"35334861-8e16-4b16-8cf5-167a049836af\") " pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:15 crc kubenswrapper[4824]: I0121 11:23:15.145202 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:15 crc kubenswrapper[4824]: I0121 11:23:15.511373 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b"] Jan 21 11:23:15 crc kubenswrapper[4824]: I0121 11:23:15.635716 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" event={"ID":"35334861-8e16-4b16-8cf5-167a049836af","Type":"ContainerStarted","Data":"f53867a58b3b691c841b96163206155dfec28ba34222d55889b813a2952b7204"} Jan 21 11:23:16 crc kubenswrapper[4824]: I0121 11:23:16.065069 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:23:16 crc kubenswrapper[4824]: I0121 11:23:16.065139 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:23:19 crc kubenswrapper[4824]: I0121 11:23:19.655701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" event={"ID":"35334861-8e16-4b16-8cf5-167a049836af","Type":"ContainerStarted","Data":"b45e8370768dd0af119e29d056ccdfde86899f4c9d7e9d40974aab7097005c2e"} Jan 21 11:23:19 crc kubenswrapper[4824]: I0121 11:23:19.656119 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:19 crc kubenswrapper[4824]: I0121 11:23:19.675920 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" podStartSLOduration=2.071477087 podStartE2EDuration="5.675906248s" podCreationTimestamp="2026-01-21 11:23:14 +0000 UTC" firstStartedPulling="2026-01-21 11:23:15.529139576 +0000 UTC m=+737.822168868" lastFinishedPulling="2026-01-21 11:23:19.133568737 +0000 UTC m=+741.426598029" observedRunningTime="2026-01-21 11:23:19.67331292 +0000 UTC m=+741.966342211" watchObservedRunningTime="2026-01-21 11:23:19.675906248 +0000 UTC m=+741.968935540" Jan 21 11:23:24 crc kubenswrapper[4824]: I0121 11:23:24.014384 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 21 11:23:25 crc kubenswrapper[4824]: I0121 11:23:25.148276 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.547309 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.548329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.549846 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-6jjxv" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.554375 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.555065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.558951 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.560576 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-2r2m7" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.567370 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.576377 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.577820 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.579023 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-w7vhc" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.582673 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.583354 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.587854 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.588776 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-vt9bz" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.595981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.608005 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.608790 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.614159 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-rpx84" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.619558 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.620265 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.624482 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-x922p" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.628416 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.634021 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636598 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h66xm\" (UniqueName: \"kubernetes.io/projected/a25733d4-a5f8-4c97-a897-8d96e637c253-kube-api-access-h66xm\") pod \"cinder-operator-controller-manager-9b68f5989-24r54\" (UID: \"a25733d4-a5f8-4c97-a897-8d96e637c253\") " pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636685 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdzth\" (UniqueName: \"kubernetes.io/projected/8ed6bb2c-2770-4785-bf38-9b6fd0bd8472-kube-api-access-vdzth\") pod \"barbican-operator-controller-manager-7ddb5c749-bvptq\" (UID: \"8ed6bb2c-2770-4785-bf38-9b6fd0bd8472\") " pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kwld\" (UniqueName: \"kubernetes.io/projected/14dd1f12-3943-42e4-be1b-c9e37e49d21b-kube-api-access-2kwld\") pod \"designate-operator-controller-manager-9f958b845-wtfj4\" (UID: \"14dd1f12-3943-42e4-be1b-c9e37e49d21b\") " pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636856 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd2tc\" (UniqueName: \"kubernetes.io/projected/944c3123-9063-4731-9581-5b683741edc4-kube-api-access-dd2tc\") pod \"heat-operator-controller-manager-5fbc59b5b6-hgmf4\" (UID: \"944c3123-9063-4731-9581-5b683741edc4\") " pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/2463afb1-0271-4ab3-ad1a-9329ecaeec4d-kube-api-access-f5dbk\") pod \"horizon-operator-controller-manager-77d5c5b54f-mp5ml\" (UID: \"2463afb1-0271-4ab3-ad1a-9329ecaeec4d\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636901 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.637055 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6kj7\" (UniqueName: \"kubernetes.io/projected/b7053a68-ce89-48b6-9b5d-1c6dac813a75-kube-api-access-h6kj7\") pod \"glance-operator-controller-manager-c6994669c-hvq6s\" (UID: \"b7053a68-ce89-48b6-9b5d-1c6dac813a75\") " pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.637649 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.640018 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-ldsjj" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.640131 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.646053 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.651557 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.652215 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.656012 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-tz46k" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.660035 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.699765 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.700613 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.704333 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-8pwkx" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.713429 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.731761 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.732507 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.735565 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.739218 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-p2zqb" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.739550 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740117 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740813 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6kj7\" (UniqueName: \"kubernetes.io/projected/b7053a68-ce89-48b6-9b5d-1c6dac813a75-kube-api-access-h6kj7\") pod \"glance-operator-controller-manager-c6994669c-hvq6s\" (UID: \"b7053a68-ce89-48b6-9b5d-1c6dac813a75\") " pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740841 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h66xm\" (UniqueName: \"kubernetes.io/projected/a25733d4-a5f8-4c97-a897-8d96e637c253-kube-api-access-h66xm\") pod \"cinder-operator-controller-manager-9b68f5989-24r54\" (UID: \"a25733d4-a5f8-4c97-a897-8d96e637c253\") " pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdzth\" (UniqueName: \"kubernetes.io/projected/8ed6bb2c-2770-4785-bf38-9b6fd0bd8472-kube-api-access-vdzth\") pod \"barbican-operator-controller-manager-7ddb5c749-bvptq\" (UID: \"8ed6bb2c-2770-4785-bf38-9b6fd0bd8472\") " pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740915 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kwld\" (UniqueName: \"kubernetes.io/projected/14dd1f12-3943-42e4-be1b-c9e37e49d21b-kube-api-access-2kwld\") pod \"designate-operator-controller-manager-9f958b845-wtfj4\" (UID: \"14dd1f12-3943-42e4-be1b-c9e37e49d21b\") " pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd2tc\" (UniqueName: \"kubernetes.io/projected/944c3123-9063-4731-9581-5b683741edc4-kube-api-access-dd2tc\") pod \"heat-operator-controller-manager-5fbc59b5b6-hgmf4\" (UID: \"944c3123-9063-4731-9581-5b683741edc4\") " pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/2463afb1-0271-4ab3-ad1a-9329ecaeec4d-kube-api-access-f5dbk\") pod \"horizon-operator-controller-manager-77d5c5b54f-mp5ml\" (UID: \"2463afb1-0271-4ab3-ad1a-9329ecaeec4d\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.748730 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-vdvrx" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.750921 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.765378 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.766091 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.767507 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-fd68l" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.774223 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kwld\" (UniqueName: \"kubernetes.io/projected/14dd1f12-3943-42e4-be1b-c9e37e49d21b-kube-api-access-2kwld\") pod \"designate-operator-controller-manager-9f958b845-wtfj4\" (UID: \"14dd1f12-3943-42e4-be1b-c9e37e49d21b\") " pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.775041 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.775881 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdzth\" (UniqueName: \"kubernetes.io/projected/8ed6bb2c-2770-4785-bf38-9b6fd0bd8472-kube-api-access-vdzth\") pod \"barbican-operator-controller-manager-7ddb5c749-bvptq\" (UID: \"8ed6bb2c-2770-4785-bf38-9b6fd0bd8472\") " pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.778674 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.780313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.781289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd2tc\" (UniqueName: \"kubernetes.io/projected/944c3123-9063-4731-9581-5b683741edc4-kube-api-access-dd2tc\") pod \"heat-operator-controller-manager-5fbc59b5b6-hgmf4\" (UID: \"944c3123-9063-4731-9581-5b683741edc4\") " pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.781532 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/2463afb1-0271-4ab3-ad1a-9329ecaeec4d-kube-api-access-f5dbk\") pod \"horizon-operator-controller-manager-77d5c5b54f-mp5ml\" (UID: \"2463afb1-0271-4ab3-ad1a-9329ecaeec4d\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.786267 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6kj7\" (UniqueName: \"kubernetes.io/projected/b7053a68-ce89-48b6-9b5d-1c6dac813a75-kube-api-access-h6kj7\") pod \"glance-operator-controller-manager-c6994669c-hvq6s\" (UID: \"b7053a68-ce89-48b6-9b5d-1c6dac813a75\") " pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.787260 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h66xm\" (UniqueName: \"kubernetes.io/projected/a25733d4-a5f8-4c97-a897-8d96e637c253-kube-api-access-h66xm\") pod \"cinder-operator-controller-manager-9b68f5989-24r54\" (UID: \"a25733d4-a5f8-4c97-a897-8d96e637c253\") " pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.789253 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-r9fdm" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.789575 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.792249 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.792879 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.794747 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-89gkm" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.797175 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.797900 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.799818 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.800026 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-ct526" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.808498 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.809940 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.810797 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.811100 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-sz9vr" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.813749 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.816747 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.822838 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.823534 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.824571 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-vlz5r" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.830883 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79prb\" (UniqueName: \"kubernetes.io/projected/19b1c150-f5ed-4432-bfc2-717b78eebc7e-kube-api-access-79prb\") pod \"mariadb-operator-controller-manager-c87fff755-d5hn9\" (UID: \"19b1c150-f5ed-4432-bfc2-717b78eebc7e\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26h5f\" (UniqueName: \"kubernetes.io/projected/2e5efa9b-40bc-4860-967d-aa37e1838f81-kube-api-access-26h5f\") pod \"keystone-operator-controller-manager-767fdc4f47-5mhmd\" (UID: \"2e5efa9b-40bc-4860-967d-aa37e1838f81\") " pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841687 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzdsw\" (UniqueName: \"kubernetes.io/projected/fc911701-7c15-4a20-b89e-709521bdd3a1-kube-api-access-qzdsw\") pod \"ironic-operator-controller-manager-78757b4889-gkwjs\" (UID: \"fc911701-7c15-4a20-b89e-709521bdd3a1\") " pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2vk8\" (UniqueName: \"kubernetes.io/projected/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-kube-api-access-r2vk8\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841767 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841792 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl8hm\" (UniqueName: \"kubernetes.io/projected/18017a19-1676-4fe9-a97e-ae878f423847-kube-api-access-vl8hm\") pod \"manila-operator-controller-manager-864f6b75bf-c7vsf\" (UID: \"18017a19-1676-4fe9-a97e-ae878f423847\") " pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.845030 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.845599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.848619 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bcx55" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.855628 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.865378 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.874555 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.893343 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.899616 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.922210 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.932517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv9nq\" (UniqueName: \"kubernetes.io/projected/8c74d3da-d870-45e9-96b9-70eb2395c977-kube-api-access-mv9nq\") pod \"neutron-operator-controller-manager-cb4666565-7w8p4\" (UID: \"8c74d3da-d870-45e9-96b9-70eb2395c977\") " pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944262 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzdsw\" (UniqueName: \"kubernetes.io/projected/fc911701-7c15-4a20-b89e-709521bdd3a1-kube-api-access-qzdsw\") pod \"ironic-operator-controller-manager-78757b4889-gkwjs\" (UID: \"fc911701-7c15-4a20-b89e-709521bdd3a1\") " pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2vk8\" (UniqueName: \"kubernetes.io/projected/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-kube-api-access-r2vk8\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944371 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl8hm\" (UniqueName: \"kubernetes.io/projected/18017a19-1676-4fe9-a97e-ae878f423847-kube-api-access-vl8hm\") pod \"manila-operator-controller-manager-864f6b75bf-c7vsf\" (UID: \"18017a19-1676-4fe9-a97e-ae878f423847\") " pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944446 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2gng\" (UniqueName: \"kubernetes.io/projected/f5639f17-3a5b-4ef4-bd6e-6eeb3274636a-kube-api-access-t2gng\") pod \"ovn-operator-controller-manager-55db956ddc-78z2q\" (UID: \"f5639f17-3a5b-4ef4-bd6e-6eeb3274636a\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79prb\" (UniqueName: \"kubernetes.io/projected/19b1c150-f5ed-4432-bfc2-717b78eebc7e-kube-api-access-79prb\") pod \"mariadb-operator-controller-manager-c87fff755-d5hn9\" (UID: \"19b1c150-f5ed-4432-bfc2-717b78eebc7e\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcm52\" (UniqueName: \"kubernetes.io/projected/3fa95966-3147-4e50-8ed8-f702a87992c4-kube-api-access-xcm52\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944534 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd9q4\" (UniqueName: \"kubernetes.io/projected/b38ac60b-1643-44d9-a1fe-5f75129247dc-kube-api-access-hd9q4\") pod \"nova-operator-controller-manager-65849867d6-8ngx6\" (UID: \"b38ac60b-1643-44d9-a1fe-5f75129247dc\") " pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944555 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk5bv\" (UniqueName: \"kubernetes.io/projected/d8b36106-8f21-4873-890d-c1879ea49068-kube-api-access-dk5bv\") pod \"placement-operator-controller-manager-686df47fcb-c8xvs\" (UID: \"d8b36106-8f21-4873-890d-c1879ea49068\") " pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944579 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944621 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26h5f\" (UniqueName: \"kubernetes.io/projected/2e5efa9b-40bc-4860-967d-aa37e1838f81-kube-api-access-26h5f\") pod \"keystone-operator-controller-manager-767fdc4f47-5mhmd\" (UID: \"2e5efa9b-40bc-4860-967d-aa37e1838f81\") " pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwdw8\" (UniqueName: \"kubernetes.io/projected/5429da27-a618-44da-8aeb-12b917af8734-kube-api-access-hwdw8\") pod \"octavia-operator-controller-manager-7fc9b76cf6-tncbx\" (UID: \"5429da27-a618-44da-8aeb-12b917af8734\") " pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.945278 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h"] Jan 21 11:23:43 crc kubenswrapper[4824]: E0121 11:23:43.945542 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:43 crc kubenswrapper[4824]: E0121 11:23:43.945582 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert podName:57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e nodeName:}" failed. No retries permitted until 2026-01-21 11:23:44.445568859 +0000 UTC m=+766.738598142 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert") pod "infra-operator-controller-manager-77c48c7859-dzn7w" (UID: "57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e") : secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.948499 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.951852 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-dptjq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.970174 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzdsw\" (UniqueName: \"kubernetes.io/projected/fc911701-7c15-4a20-b89e-709521bdd3a1-kube-api-access-qzdsw\") pod \"ironic-operator-controller-manager-78757b4889-gkwjs\" (UID: \"fc911701-7c15-4a20-b89e-709521bdd3a1\") " pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.971407 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79prb\" (UniqueName: \"kubernetes.io/projected/19b1c150-f5ed-4432-bfc2-717b78eebc7e-kube-api-access-79prb\") pod \"mariadb-operator-controller-manager-c87fff755-d5hn9\" (UID: \"19b1c150-f5ed-4432-bfc2-717b78eebc7e\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.972329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl8hm\" (UniqueName: \"kubernetes.io/projected/18017a19-1676-4fe9-a97e-ae878f423847-kube-api-access-vl8hm\") pod \"manila-operator-controller-manager-864f6b75bf-c7vsf\" (UID: \"18017a19-1676-4fe9-a97e-ae878f423847\") " pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.975797 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.976059 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2vk8\" (UniqueName: \"kubernetes.io/projected/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-kube-api-access-r2vk8\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.976272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26h5f\" (UniqueName: \"kubernetes.io/projected/2e5efa9b-40bc-4860-967d-aa37e1838f81-kube-api-access-26h5f\") pod \"keystone-operator-controller-manager-767fdc4f47-5mhmd\" (UID: \"2e5efa9b-40bc-4860-967d-aa37e1838f81\") " pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.977377 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.022128 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.046970 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.052933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2gng\" (UniqueName: \"kubernetes.io/projected/f5639f17-3a5b-4ef4-bd6e-6eeb3274636a-kube-api-access-t2gng\") pod \"ovn-operator-controller-manager-55db956ddc-78z2q\" (UID: \"f5639f17-3a5b-4ef4-bd6e-6eeb3274636a\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcm52\" (UniqueName: \"kubernetes.io/projected/3fa95966-3147-4e50-8ed8-f702a87992c4-kube-api-access-xcm52\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053171 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd9q4\" (UniqueName: \"kubernetes.io/projected/b38ac60b-1643-44d9-a1fe-5f75129247dc-kube-api-access-hd9q4\") pod \"nova-operator-controller-manager-65849867d6-8ngx6\" (UID: \"b38ac60b-1643-44d9-a1fe-5f75129247dc\") " pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk5bv\" (UniqueName: \"kubernetes.io/projected/d8b36106-8f21-4873-890d-c1879ea49068-kube-api-access-dk5bv\") pod \"placement-operator-controller-manager-686df47fcb-c8xvs\" (UID: \"d8b36106-8f21-4873-890d-c1879ea49068\") " pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053389 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwdw8\" (UniqueName: \"kubernetes.io/projected/5429da27-a618-44da-8aeb-12b917af8734-kube-api-access-hwdw8\") pod \"octavia-operator-controller-manager-7fc9b76cf6-tncbx\" (UID: \"5429da27-a618-44da-8aeb-12b917af8734\") " pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvmv5\" (UniqueName: \"kubernetes.io/projected/db84b172-59d8-4995-815b-50b492283bee-kube-api-access-jvmv5\") pod \"telemetry-operator-controller-manager-5f8f495fcf-6rz5h\" (UID: \"db84b172-59d8-4995-815b-50b492283bee\") " pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv9nq\" (UniqueName: \"kubernetes.io/projected/8c74d3da-d870-45e9-96b9-70eb2395c977-kube-api-access-mv9nq\") pod \"neutron-operator-controller-manager-cb4666565-7w8p4\" (UID: \"8c74d3da-d870-45e9-96b9-70eb2395c977\") " pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.054576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqcnj\" (UniqueName: \"kubernetes.io/projected/b4cb80e6-c720-4875-9848-7d9597a354e4-kube-api-access-kqcnj\") pod \"swift-operator-controller-manager-85dd56d4cc-wlrdt\" (UID: \"b4cb80e6-c720-4875-9848-7d9597a354e4\") " pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.057043 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.057401 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.057637 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert podName:3fa95966-3147-4e50-8ed8-f702a87992c4 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:44.557617736 +0000 UTC m=+766.850647028 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" (UID: "3fa95966-3147-4e50-8ed8-f702a87992c4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.072097 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bdjsk" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.073599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.083270 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.099075 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk5bv\" (UniqueName: \"kubernetes.io/projected/d8b36106-8f21-4873-890d-c1879ea49068-kube-api-access-dk5bv\") pod \"placement-operator-controller-manager-686df47fcb-c8xvs\" (UID: \"d8b36106-8f21-4873-890d-c1879ea49068\") " pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.100307 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd9q4\" (UniqueName: \"kubernetes.io/projected/b38ac60b-1643-44d9-a1fe-5f75129247dc-kube-api-access-hd9q4\") pod \"nova-operator-controller-manager-65849867d6-8ngx6\" (UID: \"b38ac60b-1643-44d9-a1fe-5f75129247dc\") " pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.100739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2gng\" (UniqueName: \"kubernetes.io/projected/f5639f17-3a5b-4ef4-bd6e-6eeb3274636a-kube-api-access-t2gng\") pod \"ovn-operator-controller-manager-55db956ddc-78z2q\" (UID: \"f5639f17-3a5b-4ef4-bd6e-6eeb3274636a\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.101441 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwdw8\" (UniqueName: \"kubernetes.io/projected/5429da27-a618-44da-8aeb-12b917af8734-kube-api-access-hwdw8\") pod \"octavia-operator-controller-manager-7fc9b76cf6-tncbx\" (UID: \"5429da27-a618-44da-8aeb-12b917af8734\") " pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.112429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv9nq\" (UniqueName: \"kubernetes.io/projected/8c74d3da-d870-45e9-96b9-70eb2395c977-kube-api-access-mv9nq\") pod \"neutron-operator-controller-manager-cb4666565-7w8p4\" (UID: \"8c74d3da-d870-45e9-96b9-70eb2395c977\") " pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.113526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcm52\" (UniqueName: \"kubernetes.io/projected/3fa95966-3147-4e50-8ed8-f702a87992c4-kube-api-access-xcm52\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.124537 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.125813 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.132424 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.143256 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.155410 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.156230 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.157526 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvmv5\" (UniqueName: \"kubernetes.io/projected/db84b172-59d8-4995-815b-50b492283bee-kube-api-access-jvmv5\") pod \"telemetry-operator-controller-manager-5f8f495fcf-6rz5h\" (UID: \"db84b172-59d8-4995-815b-50b492283bee\") " pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.157612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqcnj\" (UniqueName: \"kubernetes.io/projected/b4cb80e6-c720-4875-9848-7d9597a354e4-kube-api-access-kqcnj\") pod \"swift-operator-controller-manager-85dd56d4cc-wlrdt\" (UID: \"b4cb80e6-c720-4875-9848-7d9597a354e4\") " pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.157784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4547\" (UniqueName: \"kubernetes.io/projected/52491e7b-cd8c-403a-942c-312debd6f1c6-kube-api-access-r4547\") pod \"test-operator-controller-manager-7cd8bc9dbb-xkp8x\" (UID: \"52491e7b-cd8c-403a-942c-312debd6f1c6\") " pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.160777 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-qtdgh" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.161812 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.166622 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.168104 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.182179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvmv5\" (UniqueName: \"kubernetes.io/projected/db84b172-59d8-4995-815b-50b492283bee-kube-api-access-jvmv5\") pod \"telemetry-operator-controller-manager-5f8f495fcf-6rz5h\" (UID: \"db84b172-59d8-4995-815b-50b492283bee\") " pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.186506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqcnj\" (UniqueName: \"kubernetes.io/projected/b4cb80e6-c720-4875-9848-7d9597a354e4-kube-api-access-kqcnj\") pod \"swift-operator-controller-manager-85dd56d4cc-wlrdt\" (UID: \"b4cb80e6-c720-4875-9848-7d9597a354e4\") " pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.258668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4547\" (UniqueName: \"kubernetes.io/projected/52491e7b-cd8c-403a-942c-312debd6f1c6-kube-api-access-r4547\") pod \"test-operator-controller-manager-7cd8bc9dbb-xkp8x\" (UID: \"52491e7b-cd8c-403a-942c-312debd6f1c6\") " pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.258712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqc8b\" (UniqueName: \"kubernetes.io/projected/0715aa91-2ae5-4e07-8388-c37def4c7743-kube-api-access-pqc8b\") pod \"watcher-operator-controller-manager-64cd966744-9w4ch\" (UID: \"0715aa91-2ae5-4e07-8388-c37def4c7743\") " pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.283914 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.284682 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.286693 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.289232 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.295274 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.295432 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xjvvw" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.302812 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4547\" (UniqueName: \"kubernetes.io/projected/52491e7b-cd8c-403a-942c-312debd6f1c6-kube-api-access-r4547\") pod \"test-operator-controller-manager-7cd8bc9dbb-xkp8x\" (UID: \"52491e7b-cd8c-403a-942c-312debd6f1c6\") " pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.325906 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.359668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqc8b\" (UniqueName: \"kubernetes.io/projected/0715aa91-2ae5-4e07-8388-c37def4c7743-kube-api-access-pqc8b\") pod \"watcher-operator-controller-manager-64cd966744-9w4ch\" (UID: \"0715aa91-2ae5-4e07-8388-c37def4c7743\") " pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.359752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.359785 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.359867 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6spbp\" (UniqueName: \"kubernetes.io/projected/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-kube-api-access-6spbp\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.382359 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqc8b\" (UniqueName: \"kubernetes.io/projected/0715aa91-2ae5-4e07-8388-c37def4c7743-kube-api-access-pqc8b\") pod \"watcher-operator-controller-manager-64cd966744-9w4ch\" (UID: \"0715aa91-2ae5-4e07-8388-c37def4c7743\") " pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.394064 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.398702 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.399498 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.406501 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-w4k22" Jan 21 11:23:44 crc kubenswrapper[4824]: W0121 11:23:44.409735 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ed6bb2c_2770_4785_bf38_9b6fd0bd8472.slice/crio-b6219845aa9a1771a156cc32960916cee0673df44cdd949c240384b6968c356d WatchSource:0}: Error finding container b6219845aa9a1771a156cc32960916cee0673df44cdd949c240384b6968c356d: Status 404 returned error can't find the container with id b6219845aa9a1771a156cc32960916cee0673df44cdd949c240384b6968c356d Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.419163 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.428857 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.429183 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54"] Jan 21 11:23:44 crc kubenswrapper[4824]: W0121 11:23:44.434848 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda25733d4_a5f8_4c97_a897_8d96e637c253.slice/crio-f69ed3b642e0e14f2ed7c1d550fcce820181b38e087c18d787fe936ecdb84588 WatchSource:0}: Error finding container f69ed3b642e0e14f2ed7c1d550fcce820181b38e087c18d787fe936ecdb84588: Status 404 returned error can't find the container with id f69ed3b642e0e14f2ed7c1d550fcce820181b38e087c18d787fe936ecdb84588 Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.436135 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460793 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6spbp\" (UniqueName: \"kubernetes.io/projected/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-kube-api-access-6spbp\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.460804 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.460915 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:44.960902243 +0000 UTC m=+767.253931536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460867 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.460866 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.460943 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460951 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v855\" (UniqueName: \"kubernetes.io/projected/782093ab-751c-4e38-bba3-c5f76308c82e-kube-api-access-5v855\") pod \"rabbitmq-cluster-operator-manager-668c99d594-94lqx\" (UID: \"782093ab-751c-4e38-bba3-c5f76308c82e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.461026 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert podName:57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e nodeName:}" failed. No retries permitted until 2026-01-21 11:23:45.461014074 +0000 UTC m=+767.754043366 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert") pod "infra-operator-controller-manager-77c48c7859-dzn7w" (UID: "57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e") : secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.461045 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:44.96103846 +0000 UTC m=+767.254067752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.475207 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.482017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6spbp\" (UniqueName: \"kubernetes.io/projected/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-kube-api-access-6spbp\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.483570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.561675 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v855\" (UniqueName: \"kubernetes.io/projected/782093ab-751c-4e38-bba3-c5f76308c82e-kube-api-access-5v855\") pod \"rabbitmq-cluster-operator-manager-668c99d594-94lqx\" (UID: \"782093ab-751c-4e38-bba3-c5f76308c82e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.561761 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.561867 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.561903 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert podName:3fa95966-3147-4e50-8ed8-f702a87992c4 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:45.561892363 +0000 UTC m=+767.854921655 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" (UID: "3fa95966-3147-4e50-8ed8-f702a87992c4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.579542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v855\" (UniqueName: \"kubernetes.io/projected/782093ab-751c-4e38-bba3-c5f76308c82e-kube-api-access-5v855\") pod \"rabbitmq-cluster-operator-manager-668c99d594-94lqx\" (UID: \"782093ab-751c-4e38-bba3-c5f76308c82e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.639824 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.653408 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.673983 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.733448 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.775342 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.790102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" event={"ID":"2463afb1-0271-4ab3-ad1a-9329ecaeec4d","Type":"ContainerStarted","Data":"9b62e3ccc2517416d91485038728c16ef411f1e88c5143cd21730db6b19f1309"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.793571 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" event={"ID":"a25733d4-a5f8-4c97-a897-8d96e637c253","Type":"ContainerStarted","Data":"f69ed3b642e0e14f2ed7c1d550fcce820181b38e087c18d787fe936ecdb84588"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.794615 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" event={"ID":"14dd1f12-3943-42e4-be1b-c9e37e49d21b","Type":"ContainerStarted","Data":"fb96fc845fe4d6b80ac3f8cd82ccfab77cf45f6adcda11a32c14793a6e34f919"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.796725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" event={"ID":"b7053a68-ce89-48b6-9b5d-1c6dac813a75","Type":"ContainerStarted","Data":"806f3e2a06a85342fe737a02653b5b4dc955ad63bc0aa934a4f3c1f3cb9984e5"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.801161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" event={"ID":"fc911701-7c15-4a20-b89e-709521bdd3a1","Type":"ContainerStarted","Data":"fa06726926078f65ef97f8f89be9c291d385d655fc3dba940c74219d1c5360af"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.805630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" event={"ID":"8ed6bb2c-2770-4785-bf38-9b6fd0bd8472","Type":"ContainerStarted","Data":"b6219845aa9a1771a156cc32960916cee0673df44cdd949c240384b6968c356d"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.806504 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" event={"ID":"944c3123-9063-4731-9581-5b683741edc4","Type":"ContainerStarted","Data":"edc274cec7fd06e3df02f7ceb526561402e0899480a5e381bf4312fa7c43216d"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.883168 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.912005 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.932013 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.969065 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.969116 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.969290 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.969333 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:45.969320952 +0000 UTC m=+768.262350243 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.969602 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.969642 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:45.969633801 +0000 UTC m=+768.262663093 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.985892 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.988975 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.992557 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.995925 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.999080 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q"] Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.003492 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:6defa56fc6a5bfbd5b27d28ff7b1c7bc89b24b2ef956e2a6d97b2726f668a231,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hd9q4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-65849867d6-8ngx6_openstack-operators(b38ac60b-1643-44d9-a1fe-5f75129247dc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.003705 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb84b172_59d8_4995_815b_50b492283bee.slice/crio-dea14bc3dbf16dc26f2bf74918823842c7ee5b7d971485978ab0114af36bdf44 WatchSource:0}: Error finding container dea14bc3dbf16dc26f2bf74918823842c7ee5b7d971485978ab0114af36bdf44: Status 404 returned error can't find the container with id dea14bc3dbf16dc26f2bf74918823842c7ee5b7d971485978ab0114af36bdf44 Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.004677 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" podUID="b38ac60b-1643-44d9-a1fe-5f75129247dc" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.005849 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:2e89109f5db66abf1afd15ef59bda35a53db40c5e59e020579ac5aa0acea1843,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jvmv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5f8f495fcf-6rz5h_openstack-operators(db84b172-59d8-4995-815b-50b492283bee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.006979 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" podUID="db84b172-59d8-4995-815b-50b492283bee" Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.008180 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c74d3da_d870_45e9_96b9_70eb2395c977.slice/crio-cc100a8bfa18f73c4ececc9bf8a9ded089e2a55ee7ce7b6379f51ec545977f6f WatchSource:0}: Error finding container cc100a8bfa18f73c4ececc9bf8a9ded089e2a55ee7ce7b6379f51ec545977f6f: Status 404 returned error can't find the container with id cc100a8bfa18f73c4ececc9bf8a9ded089e2a55ee7ce7b6379f51ec545977f6f Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.008362 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t2gng,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-55db956ddc-78z2q_openstack-operators(f5639f17-3a5b-4ef4-bd6e-6eeb3274636a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.009581 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" podUID="f5639f17-3a5b-4ef4-bd6e-6eeb3274636a" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.010178 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mv9nq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-cb4666565-7w8p4_openstack-operators(8c74d3da-d870-45e9-96b9-70eb2395c977): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.012297 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podUID="8c74d3da-d870-45e9-96b9-70eb2395c977" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.098941 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt"] Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.103302 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch"] Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.108240 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x"] Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.110773 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4cb80e6_c720_4875_9848_7d9597a354e4.slice/crio-90eab79beadafe7c6d034fbad8cc7e89f7c0052842ddc3b3c0acf01963d3c3ab WatchSource:0}: Error finding container 90eab79beadafe7c6d034fbad8cc7e89f7c0052842ddc3b3c0acf01963d3c3ab: Status 404 returned error can't find the container with id 90eab79beadafe7c6d034fbad8cc7e89f7c0052842ddc3b3c0acf01963d3c3ab Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.112691 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52491e7b_cd8c_403a_942c_312debd6f1c6.slice/crio-512c9c60611f1470288b1906809cbd0a00165c21927d0d39083bb9fb1f9bca05 WatchSource:0}: Error finding container 512c9c60611f1470288b1906809cbd0a00165c21927d0d39083bb9fb1f9bca05: Status 404 returned error can't find the container with id 512c9c60611f1470288b1906809cbd0a00165c21927d0d39083bb9fb1f9bca05 Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.115121 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:244a4906353b84899db16a89e1ebb64491c9f85e69327cb2a72b6da0142a6e5e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r4547,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-7cd8bc9dbb-xkp8x_openstack-operators(52491e7b-cd8c-403a-942c-312debd6f1c6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.116651 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" podUID="52491e7b-cd8c-403a-942c-312debd6f1c6" Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.117576 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0715aa91_2ae5_4e07_8388_c37def4c7743.slice/crio-bbbd6d582790918c7a352ccd6df64b933c752de05fd23d6abbeeefaf76e5fd9d WatchSource:0}: Error finding container bbbd6d582790918c7a352ccd6df64b933c752de05fd23d6abbeeefaf76e5fd9d: Status 404 returned error can't find the container with id bbbd6d582790918c7a352ccd6df64b933c752de05fd23d6abbeeefaf76e5fd9d Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.122534 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d687150a46d97eb382dcd8305a2a611943af74771debe1fa9cc13a21e51c69ad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pqc8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-64cd966744-9w4ch_openstack-operators(0715aa91-2ae5-4e07-8388-c37def4c7743): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.123742 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" podUID="0715aa91-2ae5-4e07-8388-c37def4c7743" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.168530 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx"] Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.175662 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod782093ab_751c_4e38_bba3_c5f76308c82e.slice/crio-82026058c15748e5962d53b16adc2de013ffb9ac4f2a91d6edf426b7e3b96462 WatchSource:0}: Error finding container 82026058c15748e5962d53b16adc2de013ffb9ac4f2a91d6edf426b7e3b96462: Status 404 returned error can't find the container with id 82026058c15748e5962d53b16adc2de013ffb9ac4f2a91d6edf426b7e3b96462 Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.178558 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5v855,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-94lqx_openstack-operators(782093ab-751c-4e38-bba3-c5f76308c82e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.179694 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podUID="782093ab-751c-4e38-bba3-c5f76308c82e" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.476255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.476450 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.476495 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert podName:57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e nodeName:}" failed. No retries permitted until 2026-01-21 11:23:47.476480036 +0000 UTC m=+769.769509328 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert") pod "infra-operator-controller-manager-77c48c7859-dzn7w" (UID: "57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e") : secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.579863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.580052 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.580445 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert podName:3fa95966-3147-4e50-8ed8-f702a87992c4 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:47.580420518 +0000 UTC m=+769.873449810 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" (UID: "3fa95966-3147-4e50-8ed8-f702a87992c4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.830934 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" event={"ID":"782093ab-751c-4e38-bba3-c5f76308c82e","Type":"ContainerStarted","Data":"82026058c15748e5962d53b16adc2de013ffb9ac4f2a91d6edf426b7e3b96462"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.832325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" event={"ID":"d8b36106-8f21-4873-890d-c1879ea49068","Type":"ContainerStarted","Data":"db42dc90ff1dd4318f10e5f4c30a19442bab1d9aec8ed8b3ed68b96466cda44b"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.833272 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podUID="782093ab-751c-4e38-bba3-c5f76308c82e" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.836762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" event={"ID":"b38ac60b-1643-44d9-a1fe-5f75129247dc","Type":"ContainerStarted","Data":"e12efdc25c6f7aebda5d72120c94def99db956f5056672b90c777eeb9cac66d9"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.841398 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:6defa56fc6a5bfbd5b27d28ff7b1c7bc89b24b2ef956e2a6d97b2726f668a231\\\"\"" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" podUID="b38ac60b-1643-44d9-a1fe-5f75129247dc" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.841650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" event={"ID":"b4cb80e6-c720-4875-9848-7d9597a354e4","Type":"ContainerStarted","Data":"90eab79beadafe7c6d034fbad8cc7e89f7c0052842ddc3b3c0acf01963d3c3ab"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.845122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" event={"ID":"19b1c150-f5ed-4432-bfc2-717b78eebc7e","Type":"ContainerStarted","Data":"5b02d188e4a18fcba9a3114467d976279e50e34fe21b0c02722c3abf1a3e1f52"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.848989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" event={"ID":"5429da27-a618-44da-8aeb-12b917af8734","Type":"ContainerStarted","Data":"103b9e266f3a6c0be180c01b6c107e061eae09a140d6736420e12e6117cc9206"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.850385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" event={"ID":"0715aa91-2ae5-4e07-8388-c37def4c7743","Type":"ContainerStarted","Data":"bbbd6d582790918c7a352ccd6df64b933c752de05fd23d6abbeeefaf76e5fd9d"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.851994 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d687150a46d97eb382dcd8305a2a611943af74771debe1fa9cc13a21e51c69ad\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" podUID="0715aa91-2ae5-4e07-8388-c37def4c7743" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.852485 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" event={"ID":"52491e7b-cd8c-403a-942c-312debd6f1c6","Type":"ContainerStarted","Data":"512c9c60611f1470288b1906809cbd0a00165c21927d0d39083bb9fb1f9bca05"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.853465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" event={"ID":"8c74d3da-d870-45e9-96b9-70eb2395c977","Type":"ContainerStarted","Data":"cc100a8bfa18f73c4ececc9bf8a9ded089e2a55ee7ce7b6379f51ec545977f6f"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.856947 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podUID="8c74d3da-d870-45e9-96b9-70eb2395c977" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.857153 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:244a4906353b84899db16a89e1ebb64491c9f85e69327cb2a72b6da0142a6e5e\\\"\"" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" podUID="52491e7b-cd8c-403a-942c-312debd6f1c6" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.865070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" event={"ID":"2e5efa9b-40bc-4860-967d-aa37e1838f81","Type":"ContainerStarted","Data":"4c8979cc26d3eca0078286f0469ec6b079f68b32dc1981d65f9cdd1f7d5745db"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.867416 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" event={"ID":"f5639f17-3a5b-4ef4-bd6e-6eeb3274636a","Type":"ContainerStarted","Data":"1b41b5017b8d393ecb73e4a44909daf1af252b297e8f80ffd5ee58fbcb0af69b"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.868753 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" podUID="f5639f17-3a5b-4ef4-bd6e-6eeb3274636a" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.870819 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" event={"ID":"db84b172-59d8-4995-815b-50b492283bee","Type":"ContainerStarted","Data":"dea14bc3dbf16dc26f2bf74918823842c7ee5b7d971485978ab0114af36bdf44"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.872522 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:2e89109f5db66abf1afd15ef59bda35a53db40c5e59e020579ac5aa0acea1843\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" podUID="db84b172-59d8-4995-815b-50b492283bee" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.875597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" event={"ID":"18017a19-1676-4fe9-a97e-ae878f423847","Type":"ContainerStarted","Data":"126e7933e900edc5e407c8d545e0d546bd94935c101798e19e751069062d75fe"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.984940 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.985029 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.985180 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.985337 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:47.985324487 +0000 UTC m=+770.278353779 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.985196 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.985445 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:47.985428463 +0000 UTC m=+770.278457755 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.065365 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.065408 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.065470 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.065863 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.065912 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44" gracePeriod=600 Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.894386 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44" exitCode=0 Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.896208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44"} Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.898250 4824 scope.go:117] "RemoveContainer" containerID="0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.898372 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:6defa56fc6a5bfbd5b27d28ff7b1c7bc89b24b2ef956e2a6d97b2726f668a231\\\"\"" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" podUID="b38ac60b-1643-44d9-a1fe-5f75129247dc" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.899190 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podUID="8c74d3da-d870-45e9-96b9-70eb2395c977" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.909084 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" podUID="f5639f17-3a5b-4ef4-bd6e-6eeb3274636a" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.909180 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d687150a46d97eb382dcd8305a2a611943af74771debe1fa9cc13a21e51c69ad\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" podUID="0715aa91-2ae5-4e07-8388-c37def4c7743" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.909447 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:244a4906353b84899db16a89e1ebb64491c9f85e69327cb2a72b6da0142a6e5e\\\"\"" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" podUID="52491e7b-cd8c-403a-942c-312debd6f1c6" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.909508 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:2e89109f5db66abf1afd15ef59bda35a53db40c5e59e020579ac5aa0acea1843\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" podUID="db84b172-59d8-4995-815b-50b492283bee" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.910639 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podUID="782093ab-751c-4e38-bba3-c5f76308c82e" Jan 21 11:23:47 crc kubenswrapper[4824]: I0121 11:23:47.521577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:47 crc kubenswrapper[4824]: E0121 11:23:47.522046 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:47 crc kubenswrapper[4824]: E0121 11:23:47.522099 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert podName:57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e nodeName:}" failed. No retries permitted until 2026-01-21 11:23:51.522085071 +0000 UTC m=+773.815114364 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert") pod "infra-operator-controller-manager-77c48c7859-dzn7w" (UID: "57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e") : secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:47 crc kubenswrapper[4824]: I0121 11:23:47.622768 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:47 crc kubenswrapper[4824]: E0121 11:23:47.624187 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:47 crc kubenswrapper[4824]: E0121 11:23:47.624262 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert podName:3fa95966-3147-4e50-8ed8-f702a87992c4 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:51.624246459 +0000 UTC m=+773.917275752 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" (UID: "3fa95966-3147-4e50-8ed8-f702a87992c4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:48 crc kubenswrapper[4824]: I0121 11:23:48.028418 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:48 crc kubenswrapper[4824]: I0121 11:23:48.028478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:48 crc kubenswrapper[4824]: E0121 11:23:48.028565 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:23:48 crc kubenswrapper[4824]: E0121 11:23:48.028595 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:23:48 crc kubenswrapper[4824]: E0121 11:23:48.028652 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:52.02863651 +0000 UTC m=+774.321665802 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:23:48 crc kubenswrapper[4824]: E0121 11:23:48.028687 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:52.028680323 +0000 UTC m=+774.321709615 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:23:51 crc kubenswrapper[4824]: I0121 11:23:51.571932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:51 crc kubenswrapper[4824]: E0121 11:23:51.572106 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:51 crc kubenswrapper[4824]: E0121 11:23:51.572740 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert podName:57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e nodeName:}" failed. No retries permitted until 2026-01-21 11:23:59.572712107 +0000 UTC m=+781.865741399 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert") pod "infra-operator-controller-manager-77c48c7859-dzn7w" (UID: "57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e") : secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:51 crc kubenswrapper[4824]: I0121 11:23:51.673870 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:51 crc kubenswrapper[4824]: E0121 11:23:51.674058 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:51 crc kubenswrapper[4824]: E0121 11:23:51.674183 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert podName:3fa95966-3147-4e50-8ed8-f702a87992c4 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:59.67416568 +0000 UTC m=+781.967194973 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" (UID: "3fa95966-3147-4e50-8ed8-f702a87992c4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:52 crc kubenswrapper[4824]: I0121 11:23:52.079221 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:52 crc kubenswrapper[4824]: I0121 11:23:52.079280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:52 crc kubenswrapper[4824]: E0121 11:23:52.079345 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:23:52 crc kubenswrapper[4824]: E0121 11:23:52.079420 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:24:00.079404682 +0000 UTC m=+782.372433974 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:23:52 crc kubenswrapper[4824]: E0121 11:23:52.079437 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:23:52 crc kubenswrapper[4824]: E0121 11:23:52.079510 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:24:00.079495513 +0000 UTC m=+782.372524805 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:23:53 crc kubenswrapper[4824]: I0121 11:23:53.935016 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.941275 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" event={"ID":"14dd1f12-3943-42e4-be1b-c9e37e49d21b","Type":"ContainerStarted","Data":"0144a260ffe05018da1da674ab01673a24253513ff57e1761613aabaa45aa079"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.942646 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.943873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" event={"ID":"b7053a68-ce89-48b6-9b5d-1c6dac813a75","Type":"ContainerStarted","Data":"cdd57094082d6f2ee787569412482c5180a1b01f0caadb654d941d17800f292d"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.944304 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.945538 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" event={"ID":"b4cb80e6-c720-4875-9848-7d9597a354e4","Type":"ContainerStarted","Data":"b4597e6eec1e1001b39c732a9a2649db63b32895745dbe0933eff0a933c598ca"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.945902 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.952091 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" event={"ID":"19b1c150-f5ed-4432-bfc2-717b78eebc7e","Type":"ContainerStarted","Data":"01d9913b9bf0a1793eef82f6062ebc37cd27c2977db578f223abb27c36c5c629"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.952208 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.955578 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" podStartSLOduration=3.134284007 podStartE2EDuration="11.955569696s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.459904592 +0000 UTC m=+766.752933884" lastFinishedPulling="2026-01-21 11:23:53.281190282 +0000 UTC m=+775.574219573" observedRunningTime="2026-01-21 11:23:54.954534193 +0000 UTC m=+777.247563486" watchObservedRunningTime="2026-01-21 11:23:54.955569696 +0000 UTC m=+777.248598988" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.956351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" event={"ID":"5429da27-a618-44da-8aeb-12b917af8734","Type":"ContainerStarted","Data":"7413edd5c109e85ef6bac3e1732ee07186fda80dd976c15abc1f710b14a3e5d3"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.956414 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.957492 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" event={"ID":"a25733d4-a5f8-4c97-a897-8d96e637c253","Type":"ContainerStarted","Data":"fbc4fa10919d81e093dfc0ac0717ea73b800f34bdefab338d33815f6ba05d1cf"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.957914 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.959061 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" event={"ID":"fc911701-7c15-4a20-b89e-709521bdd3a1","Type":"ContainerStarted","Data":"3f749911d19fa1accdb0221d5a5e56f9cb511c2178b4f5396dcd7a3cef2360fc"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.959412 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.961026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" event={"ID":"8ed6bb2c-2770-4785-bf38-9b6fd0bd8472","Type":"ContainerStarted","Data":"c551da34bdb35bedc6b251ecd69aeb95220f8fbbd35cab2305ae053fe28ccdaa"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.961139 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.962916 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" event={"ID":"d8b36106-8f21-4873-890d-c1879ea49068","Type":"ContainerStarted","Data":"44941f67f8f6a783521e8647554e0de989687e5da65df0495c30f6faec8d8be4"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.963358 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.964830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" event={"ID":"944c3123-9063-4731-9581-5b683741edc4","Type":"ContainerStarted","Data":"d117636860c96b58d7ff7b5b600b9fa9a0115fe5fd3291c15cd27ab7fef0b75b"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.965183 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.966397 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" event={"ID":"18017a19-1676-4fe9-a97e-ae878f423847","Type":"ContainerStarted","Data":"8098b147d4a88222e038826914e86df6eb153544fedaeaa10b9c65d308469e86"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.967015 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.968533 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" event={"ID":"2463afb1-0271-4ab3-ad1a-9329ecaeec4d","Type":"ContainerStarted","Data":"7a27f2bda3a9bcde11636fb20ef32586d8b71da26ed80ee6abdef4ed31a6ef22"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.968917 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.971459 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" podStartSLOduration=2.910498731 podStartE2EDuration="11.971448419s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.68150545 +0000 UTC m=+766.974534743" lastFinishedPulling="2026-01-21 11:23:53.742455139 +0000 UTC m=+776.035484431" observedRunningTime="2026-01-21 11:23:54.966301076 +0000 UTC m=+777.259330369" watchObservedRunningTime="2026-01-21 11:23:54.971448419 +0000 UTC m=+777.264477711" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.976996 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" event={"ID":"2e5efa9b-40bc-4860-967d-aa37e1838f81","Type":"ContainerStarted","Data":"28d6b089a1084e2e9c3b6649718b08f0745f019a6d9501a71029677fabfb5b08"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.977029 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.999516 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" podStartSLOduration=3.370033267 podStartE2EDuration="11.999503578s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.113891367 +0000 UTC m=+767.406920659" lastFinishedPulling="2026-01-21 11:23:53.743361678 +0000 UTC m=+776.036390970" observedRunningTime="2026-01-21 11:23:54.997885567 +0000 UTC m=+777.290914859" watchObservedRunningTime="2026-01-21 11:23:54.999503578 +0000 UTC m=+777.292532859" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.023614 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" podStartSLOduration=3.676806274 podStartE2EDuration="12.023602667s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.934582986 +0000 UTC m=+767.227612277" lastFinishedPulling="2026-01-21 11:23:53.281379378 +0000 UTC m=+775.574408670" observedRunningTime="2026-01-21 11:23:55.019018385 +0000 UTC m=+777.312047677" watchObservedRunningTime="2026-01-21 11:23:55.023602667 +0000 UTC m=+777.316631959" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.041738 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" podStartSLOduration=3.705731862 podStartE2EDuration="12.041709461s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.530070946 +0000 UTC m=+766.823100237" lastFinishedPulling="2026-01-21 11:23:52.866048543 +0000 UTC m=+775.159077836" observedRunningTime="2026-01-21 11:23:55.038025376 +0000 UTC m=+777.331054668" watchObservedRunningTime="2026-01-21 11:23:55.041709461 +0000 UTC m=+777.334738753" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.060681 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" podStartSLOduration=3.244062133 podStartE2EDuration="12.060669373s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.931192154 +0000 UTC m=+767.224221446" lastFinishedPulling="2026-01-21 11:23:53.747799394 +0000 UTC m=+776.040828686" observedRunningTime="2026-01-21 11:23:55.056754693 +0000 UTC m=+777.349783985" watchObservedRunningTime="2026-01-21 11:23:55.060669373 +0000 UTC m=+777.353698666" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.081815 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" podStartSLOduration=3.28721898 podStartE2EDuration="12.08180101s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.000416381 +0000 UTC m=+767.293445683" lastFinishedPulling="2026-01-21 11:23:53.79499842 +0000 UTC m=+776.088027713" observedRunningTime="2026-01-21 11:23:55.077589991 +0000 UTC m=+777.370619283" watchObservedRunningTime="2026-01-21 11:23:55.08180101 +0000 UTC m=+777.374830302" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.101536 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" podStartSLOduration=3.257009882 podStartE2EDuration="12.101523719s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.897869365 +0000 UTC m=+767.190898658" lastFinishedPulling="2026-01-21 11:23:53.742383203 +0000 UTC m=+776.035412495" observedRunningTime="2026-01-21 11:23:55.097224364 +0000 UTC m=+777.390253656" watchObservedRunningTime="2026-01-21 11:23:55.101523719 +0000 UTC m=+777.394553011" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.115423 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" podStartSLOduration=3.049022588 podStartE2EDuration="12.115412041s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.675498967 +0000 UTC m=+766.968528260" lastFinishedPulling="2026-01-21 11:23:53.741888421 +0000 UTC m=+776.034917713" observedRunningTime="2026-01-21 11:23:55.111886905 +0000 UTC m=+777.404916197" watchObservedRunningTime="2026-01-21 11:23:55.115412041 +0000 UTC m=+777.408441332" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.130841 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" podStartSLOduration=2.817660107 podStartE2EDuration="12.13082743s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.428566245 +0000 UTC m=+766.721595538" lastFinishedPulling="2026-01-21 11:23:53.741733569 +0000 UTC m=+776.034762861" observedRunningTime="2026-01-21 11:23:55.129733519 +0000 UTC m=+777.422762810" watchObservedRunningTime="2026-01-21 11:23:55.13082743 +0000 UTC m=+777.423856723" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.141927 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" podStartSLOduration=3.741001391 podStartE2EDuration="12.141915312s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.453782671 +0000 UTC m=+766.746811964" lastFinishedPulling="2026-01-21 11:23:52.854696593 +0000 UTC m=+775.147725885" observedRunningTime="2026-01-21 11:23:55.141682444 +0000 UTC m=+777.434711735" watchObservedRunningTime="2026-01-21 11:23:55.141915312 +0000 UTC m=+777.434944604" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.160906 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" podStartSLOduration=3.032802018 podStartE2EDuration="12.16089421s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.675232235 +0000 UTC m=+766.968261516" lastFinishedPulling="2026-01-21 11:23:53.803324416 +0000 UTC m=+776.096353708" observedRunningTime="2026-01-21 11:23:55.160131944 +0000 UTC m=+777.453161235" watchObservedRunningTime="2026-01-21 11:23:55.16089421 +0000 UTC m=+777.453923502" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.173347 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" podStartSLOduration=3.207844527 podStartE2EDuration="12.173335624s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.79080212 +0000 UTC m=+767.083831402" lastFinishedPulling="2026-01-21 11:23:53.756293206 +0000 UTC m=+776.049322499" observedRunningTime="2026-01-21 11:23:55.17291466 +0000 UTC m=+777.465943962" watchObservedRunningTime="2026-01-21 11:23:55.173335624 +0000 UTC m=+777.466364916" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.584329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.589251 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.686198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.689586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.753970 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-ct526" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.763003 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.858989 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-ldsjj" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.867169 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.004369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" event={"ID":"db84b172-59d8-4995-815b-50b492283bee","Type":"ContainerStarted","Data":"d2c19ea693d44ba91ea907492ee970b2352043e1ba248842c926563c68d5ba97"} Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.004767 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.020062 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" podStartSLOduration=2.593054925 podStartE2EDuration="17.020049241s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.005765555 +0000 UTC m=+767.298794847" lastFinishedPulling="2026-01-21 11:23:59.432759872 +0000 UTC m=+781.725789163" observedRunningTime="2026-01-21 11:24:00.015951146 +0000 UTC m=+782.308980438" watchObservedRunningTime="2026-01-21 11:24:00.020049241 +0000 UTC m=+782.313078533" Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.090475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.090530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:00 crc kubenswrapper[4824]: E0121 11:24:00.090631 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:24:00 crc kubenswrapper[4824]: E0121 11:24:00.090688 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:24:00 crc kubenswrapper[4824]: E0121 11:24:00.090698 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:24:16.090667363 +0000 UTC m=+798.383696655 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:24:00 crc kubenswrapper[4824]: E0121 11:24:00.090798 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:24:16.090766269 +0000 UTC m=+798.383795561 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.134732 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl"] Jan 21 11:24:00 crc kubenswrapper[4824]: W0121 11:24:00.140728 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fa95966_3147_4e50_8ed8_f702a87992c4.slice/crio-eed19a8e42f8cee6fa76d0869d23edca90ebfc224ff7958bd7850ad2a22a3286 WatchSource:0}: Error finding container eed19a8e42f8cee6fa76d0869d23edca90ebfc224ff7958bd7850ad2a22a3286: Status 404 returned error can't find the container with id eed19a8e42f8cee6fa76d0869d23edca90ebfc224ff7958bd7850ad2a22a3286 Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.228752 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w"] Jan 21 11:24:00 crc kubenswrapper[4824]: W0121 11:24:00.232626 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57e1fe13_b2c8_4711_9efe_1ebbc7e9d12e.slice/crio-1fdcea483970f40088fb49b697cb5f962853a54895fc1319cad82d4dfde02a2f WatchSource:0}: Error finding container 1fdcea483970f40088fb49b697cb5f962853a54895fc1319cad82d4dfde02a2f: Status 404 returned error can't find the container with id 1fdcea483970f40088fb49b697cb5f962853a54895fc1319cad82d4dfde02a2f Jan 21 11:24:01 crc kubenswrapper[4824]: I0121 11:24:01.014197 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" event={"ID":"3fa95966-3147-4e50-8ed8-f702a87992c4","Type":"ContainerStarted","Data":"eed19a8e42f8cee6fa76d0869d23edca90ebfc224ff7958bd7850ad2a22a3286"} Jan 21 11:24:01 crc kubenswrapper[4824]: I0121 11:24:01.017881 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" event={"ID":"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e","Type":"ContainerStarted","Data":"1fdcea483970f40088fb49b697cb5f962853a54895fc1319cad82d4dfde02a2f"} Jan 21 11:24:01 crc kubenswrapper[4824]: I0121 11:24:01.019892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" event={"ID":"b38ac60b-1643-44d9-a1fe-5f75129247dc","Type":"ContainerStarted","Data":"dab60b48ed9b2e82be6ad6709abeac7a43f8f3c164163b277dd4c72e10eff626"} Jan 21 11:24:01 crc kubenswrapper[4824]: I0121 11:24:01.021433 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:24:01 crc kubenswrapper[4824]: I0121 11:24:01.034045 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" podStartSLOduration=2.377611408 podStartE2EDuration="18.034029617s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.003380289 +0000 UTC m=+767.296409581" lastFinishedPulling="2026-01-21 11:24:00.659798498 +0000 UTC m=+782.952827790" observedRunningTime="2026-01-21 11:24:01.033572825 +0000 UTC m=+783.326602108" watchObservedRunningTime="2026-01-21 11:24:01.034029617 +0000 UTC m=+783.327058908" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.039127 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" event={"ID":"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e","Type":"ContainerStarted","Data":"6f712b0e06905ad53625e13daa538506c186e3d2f91f50bf50f6d5e64d5de21d"} Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.040287 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.043709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" event={"ID":"f5639f17-3a5b-4ef4-bd6e-6eeb3274636a","Type":"ContainerStarted","Data":"88e327fb8ddda307734a971cf283fd5be51c0aecdfa49eb1437c82a96f083bf9"} Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.044093 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.044981 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" event={"ID":"3fa95966-3147-4e50-8ed8-f702a87992c4","Type":"ContainerStarted","Data":"4b62ac54cccbebffb3336a2120bc91a3453bdc07a6c7b190cbcaa9142b158d2e"} Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.045304 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.057305 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" podStartSLOduration=17.588435571 podStartE2EDuration="20.057294877s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:24:00.234661875 +0000 UTC m=+782.527691167" lastFinishedPulling="2026-01-21 11:24:02.70352118 +0000 UTC m=+784.996550473" observedRunningTime="2026-01-21 11:24:03.05209682 +0000 UTC m=+785.345126111" watchObservedRunningTime="2026-01-21 11:24:03.057294877 +0000 UTC m=+785.350324169" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.075226 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" podStartSLOduration=2.380820997 podStartE2EDuration="20.075210496s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.008271108 +0000 UTC m=+767.301300401" lastFinishedPulling="2026-01-21 11:24:02.702660608 +0000 UTC m=+784.995689900" observedRunningTime="2026-01-21 11:24:03.064194763 +0000 UTC m=+785.357224055" watchObservedRunningTime="2026-01-21 11:24:03.075210496 +0000 UTC m=+785.368239789" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.090178 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" podStartSLOduration=17.509007375 podStartE2EDuration="20.090159786s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:24:00.145979694 +0000 UTC m=+782.439008986" lastFinishedPulling="2026-01-21 11:24:02.727132105 +0000 UTC m=+785.020161397" observedRunningTime="2026-01-21 11:24:03.084979351 +0000 UTC m=+785.378008643" watchObservedRunningTime="2026-01-21 11:24:03.090159786 +0000 UTC m=+785.383189078" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.868645 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.878531 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.896648 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.906583 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.925027 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.935606 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.978655 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.024109 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.077535 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.085784 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.146274 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.172459 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.293915 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.483297 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:24:05 crc kubenswrapper[4824]: I0121 11:24:05.057516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" event={"ID":"52491e7b-cd8c-403a-942c-312debd6f1c6","Type":"ContainerStarted","Data":"1dc24aac38479c7e94c84527a6d3167ce426dc8f5e9c76700dd730953f534c97"} Jan 21 11:24:05 crc kubenswrapper[4824]: I0121 11:24:05.057647 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:24:05 crc kubenswrapper[4824]: I0121 11:24:05.058811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" event={"ID":"0715aa91-2ae5-4e07-8388-c37def4c7743","Type":"ContainerStarted","Data":"1f098a0e7196ad5933f6d0820f01f162fefc145fc5caad398e23c286893eae93"} Jan 21 11:24:05 crc kubenswrapper[4824]: I0121 11:24:05.059110 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:24:05 crc kubenswrapper[4824]: I0121 11:24:05.072154 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" podStartSLOduration=1.7818485069999999 podStartE2EDuration="21.072138393s" podCreationTimestamp="2026-01-21 11:23:44 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.115013422 +0000 UTC m=+767.408042715" lastFinishedPulling="2026-01-21 11:24:04.40530331 +0000 UTC m=+786.698332601" observedRunningTime="2026-01-21 11:24:05.068497329 +0000 UTC m=+787.361526611" watchObservedRunningTime="2026-01-21 11:24:05.072138393 +0000 UTC m=+787.365167685" Jan 21 11:24:09 crc kubenswrapper[4824]: E0121 11:24:09.500904 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/1d/1dd2737718dec9f0ef230eebb751bd30ae264b25cf038f0f2c5932675c31963c?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112359Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=2908eeb17e9386508e401967adadfeeb1a8c24ec763c4f2a1f640fe7038c9d90®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=rabbitmq-cluster-operator&akamai_signature=exp=1768995539~hmac=f620f0ae3fbab2bcd94414a12289b90d9494338380b60c29155025ae5b7b0e0b\": net/http: TLS handshake timeout" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Jan 21 11:24:09 crc kubenswrapper[4824]: E0121 11:24:09.501501 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5v855,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-94lqx_openstack-operators(782093ab-751c-4e38-bba3-c5f76308c82e): ErrImagePull: parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/1d/1dd2737718dec9f0ef230eebb751bd30ae264b25cf038f0f2c5932675c31963c?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112359Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=2908eeb17e9386508e401967adadfeeb1a8c24ec763c4f2a1f640fe7038c9d90®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=rabbitmq-cluster-operator&akamai_signature=exp=1768995539~hmac=f620f0ae3fbab2bcd94414a12289b90d9494338380b60c29155025ae5b7b0e0b\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 21 11:24:09 crc kubenswrapper[4824]: E0121 11:24:09.502707 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"parsing image configuration: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/1d/1dd2737718dec9f0ef230eebb751bd30ae264b25cf038f0f2c5932675c31963c?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112359Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=2908eeb17e9386508e401967adadfeeb1a8c24ec763c4f2a1f640fe7038c9d90®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=rabbitmq-cluster-operator&akamai_signature=exp=1768995539~hmac=f620f0ae3fbab2bcd94414a12289b90d9494338380b60c29155025ae5b7b0e0b\\\": net/http: TLS handshake timeout\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podUID="782093ab-751c-4e38-bba3-c5f76308c82e" Jan 21 11:24:09 crc kubenswrapper[4824]: I0121 11:24:09.769659 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:24:09 crc kubenswrapper[4824]: I0121 11:24:09.790027 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" podStartSLOduration=6.50495367 podStartE2EDuration="25.790012526s" podCreationTimestamp="2026-01-21 11:23:44 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.122431636 +0000 UTC m=+767.415460929" lastFinishedPulling="2026-01-21 11:24:04.407490493 +0000 UTC m=+786.700519785" observedRunningTime="2026-01-21 11:24:05.087530425 +0000 UTC m=+787.380559717" watchObservedRunningTime="2026-01-21 11:24:09.790012526 +0000 UTC m=+792.083041818" Jan 21 11:24:09 crc kubenswrapper[4824]: I0121 11:24:09.872704 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:24:11 crc kubenswrapper[4824]: E0121 11:24:11.502498 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/af/afdc062de42921db29b4b435d761747aae5b8aa0eceee742dd2e647442dbf521?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112401Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=837dd8e842ca293dd58f711fd3a985c4f6012b6b591e06d0baa4bdc0876561ac®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=neutron-operator&akamai_signature=exp=1768995541~hmac=2f4c7f6edf21ae166ff1bf15ce0fa3c5672c3d4860a1962c6454928ba7c14a2c\": net/http: TLS handshake timeout" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c" Jan 21 11:24:11 crc kubenswrapper[4824]: E0121 11:24:11.503025 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mv9nq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-cb4666565-7w8p4_openstack-operators(8c74d3da-d870-45e9-96b9-70eb2395c977): ErrImagePull: parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/af/afdc062de42921db29b4b435d761747aae5b8aa0eceee742dd2e647442dbf521?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112401Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=837dd8e842ca293dd58f711fd3a985c4f6012b6b591e06d0baa4bdc0876561ac®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=neutron-operator&akamai_signature=exp=1768995541~hmac=2f4c7f6edf21ae166ff1bf15ce0fa3c5672c3d4860a1962c6454928ba7c14a2c\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 21 11:24:11 crc kubenswrapper[4824]: E0121 11:24:11.504185 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"parsing image configuration: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/af/afdc062de42921db29b4b435d761747aae5b8aa0eceee742dd2e647442dbf521?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112401Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=837dd8e842ca293dd58f711fd3a985c4f6012b6b591e06d0baa4bdc0876561ac®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=neutron-operator&akamai_signature=exp=1768995541~hmac=2f4c7f6edf21ae166ff1bf15ce0fa3c5672c3d4860a1962c6454928ba7c14a2c\\\": net/http: TLS handshake timeout\"" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podUID="8c74d3da-d870-45e9-96b9-70eb2395c977" Jan 21 11:24:14 crc kubenswrapper[4824]: I0121 11:24:14.136419 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:24:14 crc kubenswrapper[4824]: I0121 11:24:14.164533 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:24:14 crc kubenswrapper[4824]: I0121 11:24:14.432135 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:24:14 crc kubenswrapper[4824]: I0121 11:24:14.486001 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.191950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.192268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.197197 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.197314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.476687 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xjvvw" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.485625 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.838425 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps"] Jan 21 11:24:17 crc kubenswrapper[4824]: I0121 11:24:17.133986 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" event={"ID":"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8","Type":"ContainerStarted","Data":"9d0b177e548eb2e34aa7bd44e026feaa41ad53dd5eac90d7991be6b44e6233bc"} Jan 21 11:24:17 crc kubenswrapper[4824]: I0121 11:24:17.134302 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:17 crc kubenswrapper[4824]: I0121 11:24:17.134315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" event={"ID":"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8","Type":"ContainerStarted","Data":"dfdacc7e7292e1d386dfca28ca5044d5c10632f2eb094375d8a700c43363fbf0"} Jan 21 11:24:17 crc kubenswrapper[4824]: I0121 11:24:17.154865 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" podStartSLOduration=33.154843049 podStartE2EDuration="33.154843049s" podCreationTimestamp="2026-01-21 11:23:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:24:17.153971295 +0000 UTC m=+799.447000587" watchObservedRunningTime="2026-01-21 11:24:17.154843049 +0000 UTC m=+799.447872341" Jan 21 11:24:20 crc kubenswrapper[4824]: E0121 11:24:20.050590 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podUID="782093ab-751c-4e38-bba3-c5f76308c82e" Jan 21 11:24:25 crc kubenswrapper[4824]: E0121 11:24:25.050441 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podUID="8c74d3da-d870-45e9-96b9-70eb2395c977" Jan 21 11:24:26 crc kubenswrapper[4824]: I0121 11:24:26.490457 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:39 crc kubenswrapper[4824]: I0121 11:24:39.248390 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" event={"ID":"782093ab-751c-4e38-bba3-c5f76308c82e","Type":"ContainerStarted","Data":"4c12156aea277683ef2931fbe886c4ae11e0b16b83a9ffb882ec8ab5f8cf54f7"} Jan 21 11:24:39 crc kubenswrapper[4824]: I0121 11:24:39.260734 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podStartSLOduration=2.158120752 podStartE2EDuration="55.260720716s" podCreationTimestamp="2026-01-21 11:23:44 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.178437605 +0000 UTC m=+767.471466897" lastFinishedPulling="2026-01-21 11:24:38.281037569 +0000 UTC m=+820.574066861" observedRunningTime="2026-01-21 11:24:39.259497109 +0000 UTC m=+821.552526401" watchObservedRunningTime="2026-01-21 11:24:39.260720716 +0000 UTC m=+821.553750009" Jan 21 11:24:41 crc kubenswrapper[4824]: I0121 11:24:41.258187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" event={"ID":"8c74d3da-d870-45e9-96b9-70eb2395c977","Type":"ContainerStarted","Data":"0bc7d1544e430cce051bfb7bf230c2897a71d350529bafdf4faae5c98df4f629"} Jan 21 11:24:41 crc kubenswrapper[4824]: I0121 11:24:41.258723 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:24:41 crc kubenswrapper[4824]: I0121 11:24:41.272209 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podStartSLOduration=2.704576031 podStartE2EDuration="58.272189561s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.010099676 +0000 UTC m=+767.303128968" lastFinishedPulling="2026-01-21 11:24:40.577713205 +0000 UTC m=+822.870742498" observedRunningTime="2026-01-21 11:24:41.268641813 +0000 UTC m=+823.561671105" watchObservedRunningTime="2026-01-21 11:24:41.272189561 +0000 UTC m=+823.565218853" Jan 21 11:24:54 crc kubenswrapper[4824]: I0121 11:24:54.127888 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.913985 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.915520 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.917301 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.917657 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.917676 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.917950 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-r6gf5" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.925103 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.971391 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.972421 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.974290 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.980288 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.032123 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.032181 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.032233 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt49s\" (UniqueName: \"kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.032277 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q4rz\" (UniqueName: \"kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.032301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.133566 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.133648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt49s\" (UniqueName: \"kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.133725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q4rz\" (UniqueName: \"kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.133927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.134289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.134593 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.134857 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.134985 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.148564 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q4rz\" (UniqueName: \"kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.149879 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt49s\" (UniqueName: \"kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.230986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.283816 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.613165 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.668518 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:07 crc kubenswrapper[4824]: W0121 11:25:07.672199 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8aca65b_76cd_45f9_8b20_da2e34a41fc2.slice/crio-8c626b420278544fc13bdb44a0fc27116c48e8b3b2e62a1f886c8aa8d6e59139 WatchSource:0}: Error finding container 8c626b420278544fc13bdb44a0fc27116c48e8b3b2e62a1f886c8aa8d6e59139: Status 404 returned error can't find the container with id 8c626b420278544fc13bdb44a0fc27116c48e8b3b2e62a1f886c8aa8d6e59139 Jan 21 11:25:08 crc kubenswrapper[4824]: I0121 11:25:08.385778 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" event={"ID":"b48ab67f-88cb-4921-b6a3-38b77b534bf9","Type":"ContainerStarted","Data":"c935cf67bf3bb60e71d9e8fa32ea33c10f87bb841cf838b50e9a195acf0ce5f3"} Jan 21 11:25:08 crc kubenswrapper[4824]: I0121 11:25:08.387466 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" event={"ID":"e8aca65b-76cd-45f9-8b20-da2e34a41fc2","Type":"ContainerStarted","Data":"8c626b420278544fc13bdb44a0fc27116c48e8b3b2e62a1f886c8aa8d6e59139"} Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.887148 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.911040 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.920300 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.920382 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.979839 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.979967 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.980034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msx57\" (UniqueName: \"kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.086472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msx57\" (UniqueName: \"kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.086670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.089973 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.090080 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.091478 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.111987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msx57\" (UniqueName: \"kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.154905 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.174007 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.175237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.187988 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.240471 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.295677 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.297412 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.297585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcgfn\" (UniqueName: \"kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.398886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.399094 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcgfn\" (UniqueName: \"kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.399288 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.399752 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.401449 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.416739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcgfn\" (UniqueName: \"kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.495899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.047060 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.049320 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.051741 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052103 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052321 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052468 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-22jwj" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052527 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052619 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052738 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.065722 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109008 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109049 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109108 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109280 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109313 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109343 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkwkl\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211439 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkwkl\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211571 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211665 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211767 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.212719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.213164 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.214482 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.214564 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.215373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.217217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.217309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.217363 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.224829 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.225348 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.226608 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkwkl\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.234922 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.306244 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.307358 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.309442 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.309739 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.309858 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.309932 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.309997 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.311777 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zjdwv" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.311783 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.312529 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.372007 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413539 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413667 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413786 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9nc5\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515162 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515333 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9nc5\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515457 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515547 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515995 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.516734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.517429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.518773 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.519310 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.519554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.519851 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.522441 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.530712 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9nc5\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.530938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.630706 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.647872 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.658821 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.658910 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.667322 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.668033 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.668145 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-wrlfw" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.668745 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.670499 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833708 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-config-data-default\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833733 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/218aca21-4c37-4862-b891-fc228baa5348-config-data-generated\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833774 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-kolla-config\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833816 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833861 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-operator-scripts\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkxx7\" (UniqueName: \"kubernetes.io/projected/218aca21-4c37-4862-b891-fc228baa5348-kube-api-access-bkxx7\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.934933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-operator-scripts\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.934989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkxx7\" (UniqueName: \"kubernetes.io/projected/218aca21-4c37-4862-b891-fc228baa5348-kube-api-access-bkxx7\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935029 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-config-data-default\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935123 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/218aca21-4c37-4862-b891-fc228baa5348-config-data-generated\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-kolla-config\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935203 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935439 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/218aca21-4c37-4862-b891-fc228baa5348-config-data-generated\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.937156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-config-data-default\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.937236 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-kolla-config\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.937315 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.938469 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-operator-scripts\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.948244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.950854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.961307 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkxx7\" (UniqueName: \"kubernetes.io/projected/218aca21-4c37-4862-b891-fc228baa5348-kube-api-access-bkxx7\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.965464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.975443 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.001298 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.005440 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.007247 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.007925 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.008065 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-4hq4q" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.008242 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.019421 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.054570 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt9bn\" (UniqueName: \"kubernetes.io/projected/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kube-api-access-tt9bn\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.054814 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.054861 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.054898 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.054974 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.055002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.055047 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.055067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156597 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156640 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt9bn\" (UniqueName: \"kubernetes.io/projected/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kube-api-access-tt9bn\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156796 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156821 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156845 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.157051 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.157580 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.157642 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.158020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.158798 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.160646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.160767 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.175238 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.177000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt9bn\" (UniqueName: \"kubernetes.io/projected/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kube-api-access-tt9bn\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.262654 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.266925 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.270446 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.270839 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-vxswb" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.276297 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.280582 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.324417 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.463160 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.463237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r5tc\" (UniqueName: \"kubernetes.io/projected/dba0050b-8a73-4355-a1e0-7c9a03557ead-kube-api-access-7r5tc\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.463258 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.463300 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-kolla-config\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.463405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-config-data\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.565249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.565316 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r5tc\" (UniqueName: \"kubernetes.io/projected/dba0050b-8a73-4355-a1e0-7c9a03557ead-kube-api-access-7r5tc\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.565336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.565374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-kolla-config\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.565419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-config-data\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.566099 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-config-data\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.566282 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-kolla-config\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.570352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.572123 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.580654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r5tc\" (UniqueName: \"kubernetes.io/projected/dba0050b-8a73-4355-a1e0-7c9a03557ead-kube-api-access-7r5tc\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.590706 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.106715 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.107903 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.109346 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-464bp" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.124321 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.288516 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6ndv\" (UniqueName: \"kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv\") pod \"kube-state-metrics-0\" (UID: \"2e3a943c-ee06-4e54-a984-34ecb1320f89\") " pod="openstack/kube-state-metrics-0" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.390303 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6ndv\" (UniqueName: \"kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv\") pod \"kube-state-metrics-0\" (UID: \"2e3a943c-ee06-4e54-a984-34ecb1320f89\") " pod="openstack/kube-state-metrics-0" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.408156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6ndv\" (UniqueName: \"kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv\") pod \"kube-state-metrics-0\" (UID: \"2e3a943c-ee06-4e54-a984-34ecb1320f89\") " pod="openstack/kube-state-metrics-0" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.429098 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.027448 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: W0121 11:25:19.033844 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod156fec50_b486_4e84_a7bf_b40491a863c7.slice/crio-37d36a29724a133bc6f138979451aa8f9c0eeec9cad658937121a853dcbd295e WatchSource:0}: Error finding container 37d36a29724a133bc6f138979451aa8f9c0eeec9cad658937121a853dcbd295e: Status 404 returned error can't find the container with id 37d36a29724a133bc6f138979451aa8f9c0eeec9cad658937121a853dcbd295e Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.035950 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.040221 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:19 crc kubenswrapper[4824]: W0121 11:25:19.045246 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf67dfc52_dc0e_404e_af22_75c6a3d10721.slice/crio-4c0d7f6c5bc10bc5e3cf172c7b361f178beb8f1decb24246db808a7099969563 WatchSource:0}: Error finding container 4c0d7f6c5bc10bc5e3cf172c7b361f178beb8f1decb24246db808a7099969563: Status 404 returned error can't find the container with id 4c0d7f6c5bc10bc5e3cf172c7b361f178beb8f1decb24246db808a7099969563 Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.142211 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.147392 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.226467 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.230618 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.241180 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.245725 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.461368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2e3a943c-ee06-4e54-a984-34ecb1320f89","Type":"ContainerStarted","Data":"6cb5ad5332507e249a47db656294541a580560b8688929dae90a9899b846d8f8"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.462363 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerStarted","Data":"37d36a29724a133bc6f138979451aa8f9c0eeec9cad658937121a853dcbd295e"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.463771 4824 generic.go:334] "Generic (PLEG): container finished" podID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerID="2fabf442546067371acccd858ab70f06b5ad7ea6a02d270ec355916d978e1384" exitCode=0 Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.463810 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" event={"ID":"f67dfc52-dc0e-404e-af22-75c6a3d10721","Type":"ContainerDied","Data":"2fabf442546067371acccd858ab70f06b5ad7ea6a02d270ec355916d978e1384"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.463826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" event={"ID":"f67dfc52-dc0e-404e-af22-75c6a3d10721","Type":"ContainerStarted","Data":"4c0d7f6c5bc10bc5e3cf172c7b361f178beb8f1decb24246db808a7099969563"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.465171 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerStarted","Data":"8ee895c662c7814313d66afa49854eb939a2dc25462867e35d0362794391ee25"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.466079 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"218aca21-4c37-4862-b891-fc228baa5348","Type":"ContainerStarted","Data":"bf9686d495d7804b469a74da5f65df348ba59c6886dccf9481b92b03528d8379"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.467003 4824 generic.go:334] "Generic (PLEG): container finished" podID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerID="cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059" exitCode=0 Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.467109 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" event={"ID":"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4","Type":"ContainerDied","Data":"cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.467130 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" event={"ID":"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4","Type":"ContainerStarted","Data":"9f481192efadf26388108ffc156d5c70e082a7630e4e6aa574c82729d2674bf6"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.468311 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dba0050b-8a73-4355-a1e0-7c9a03557ead","Type":"ContainerStarted","Data":"0a266bda78772602cb7070fdd028643409e691a2f47d4b4ff003042c9da515e1"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.469342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346","Type":"ContainerStarted","Data":"e38b6eb1d31e1293e699d97ffa6c84831dffe4bbdbcaad6d49e66ab12b79817c"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.470857 4824 generic.go:334] "Generic (PLEG): container finished" podID="b48ab67f-88cb-4921-b6a3-38b77b534bf9" containerID="415de3e8643b808347c13b2c7d4d57542a514312e1a5bd379cbc7b434241d117" exitCode=0 Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.470902 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" event={"ID":"b48ab67f-88cb-4921-b6a3-38b77b534bf9","Type":"ContainerDied","Data":"415de3e8643b808347c13b2c7d4d57542a514312e1a5bd379cbc7b434241d117"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.472284 4824 generic.go:334] "Generic (PLEG): container finished" podID="e8aca65b-76cd-45f9-8b20-da2e34a41fc2" containerID="aaacc227843f224a20462553c49fdbfc00aeab8ce2765ad6ee4c575c381c5973" exitCode=0 Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.472314 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" event={"ID":"e8aca65b-76cd-45f9-8b20-da2e34a41fc2","Type":"ContainerDied","Data":"aaacc227843f224a20462553c49fdbfc00aeab8ce2765ad6ee4c575c381c5973"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.775886 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-htwd6"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.776848 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.784748 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.784951 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.785106 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-ktgdl" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.798775 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-htwd6"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.831513 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-gr9gj"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.833539 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.845287 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gr9gj"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.915253 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.936183 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.956514 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-lib\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957302 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-scripts\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-log\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957389 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-combined-ca-bundle\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957412 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957445 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-ovn-controller-tls-certs\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/511f2b6d-a08d-49f8-b393-ab222219d4a7-scripts\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-run\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz8l5\" (UniqueName: \"kubernetes.io/projected/511f2b6d-a08d-49f8-b393-ab222219d4a7-kube-api-access-cz8l5\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957766 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cdxj\" (UniqueName: \"kubernetes.io/projected/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-kube-api-access-6cdxj\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957816 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-etc-ovs\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-log-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058414 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q4rz\" (UniqueName: \"kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz\") pod \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058577 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc\") pod \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058676 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config\") pod \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058736 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config\") pod \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058766 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt49s\" (UniqueName: \"kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s\") pod \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-etc-ovs\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-log-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059045 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-lib\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059065 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-scripts\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-log\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059099 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059116 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-combined-ca-bundle\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059432 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-ovn-controller-tls-certs\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-log-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.060125 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-log\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.061448 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-etc-ovs\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.061624 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/511f2b6d-a08d-49f8-b393-ab222219d4a7-scripts\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.061671 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-run\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.061702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz8l5\" (UniqueName: \"kubernetes.io/projected/511f2b6d-a08d-49f8-b393-ab222219d4a7-kube-api-access-cz8l5\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.061719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-lib\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.062162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.062232 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-run\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.063050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cdxj\" (UniqueName: \"kubernetes.io/projected/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-kube-api-access-6cdxj\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.063168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s" (OuterVolumeSpecName: "kube-api-access-jt49s") pod "e8aca65b-76cd-45f9-8b20-da2e34a41fc2" (UID: "e8aca65b-76cd-45f9-8b20-da2e34a41fc2"). InnerVolumeSpecName "kube-api-access-jt49s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.063208 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.063349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz" (OuterVolumeSpecName: "kube-api-access-8q4rz") pod "b48ab67f-88cb-4921-b6a3-38b77b534bf9" (UID: "b48ab67f-88cb-4921-b6a3-38b77b534bf9"). InnerVolumeSpecName "kube-api-access-8q4rz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.067113 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-scripts\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.067269 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-combined-ca-bundle\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.067402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/511f2b6d-a08d-49f8-b393-ab222219d4a7-scripts\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.074740 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8aca65b-76cd-45f9-8b20-da2e34a41fc2" (UID: "e8aca65b-76cd-45f9-8b20-da2e34a41fc2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.074913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-ovn-controller-tls-certs\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.075586 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config" (OuterVolumeSpecName: "config") pod "e8aca65b-76cd-45f9-8b20-da2e34a41fc2" (UID: "e8aca65b-76cd-45f9-8b20-da2e34a41fc2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.076992 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz8l5\" (UniqueName: \"kubernetes.io/projected/511f2b6d-a08d-49f8-b393-ab222219d4a7-kube-api-access-cz8l5\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.077558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cdxj\" (UniqueName: \"kubernetes.io/projected/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-kube-api-access-6cdxj\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.083243 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config" (OuterVolumeSpecName: "config") pod "b48ab67f-88cb-4921-b6a3-38b77b534bf9" (UID: "b48ab67f-88cb-4921-b6a3-38b77b534bf9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.106911 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.165464 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q4rz\" (UniqueName: \"kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.165490 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.165499 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.165507 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.165515 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt49s\" (UniqueName: \"kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.211781 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.319021 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 21 11:25:20 crc kubenswrapper[4824]: E0121 11:25:20.319284 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48ab67f-88cb-4921-b6a3-38b77b534bf9" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.319298 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48ab67f-88cb-4921-b6a3-38b77b534bf9" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: E0121 11:25:20.319316 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8aca65b-76cd-45f9-8b20-da2e34a41fc2" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.319321 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8aca65b-76cd-45f9-8b20-da2e34a41fc2" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.319452 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8aca65b-76cd-45f9-8b20-da2e34a41fc2" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.319468 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b48ab67f-88cb-4921-b6a3-38b77b534bf9" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.320122 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.322531 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.322826 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.326133 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.326259 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.326520 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-l8pwl" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.331371 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-config\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367789 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367884 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2nbk\" (UniqueName: \"kubernetes.io/projected/5c0b5876-da04-49e6-8946-4fc6060a64ec-kube-api-access-x2nbk\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367986 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.368067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471027 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-config\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471123 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471208 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471258 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2nbk\" (UniqueName: \"kubernetes.io/projected/5c0b5876-da04-49e6-8946-4fc6060a64ec-kube-api-access-x2nbk\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471327 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471599 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471982 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-config\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.472242 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.474547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.474722 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.478521 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.483830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" event={"ID":"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4","Type":"ContainerStarted","Data":"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8"} Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.483884 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.485157 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2nbk\" (UniqueName: \"kubernetes.io/projected/5c0b5876-da04-49e6-8946-4fc6060a64ec-kube-api-access-x2nbk\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.489639 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" event={"ID":"f67dfc52-dc0e-404e-af22-75c6a3d10721","Type":"ContainerStarted","Data":"0ad321acd47c8d7e7d82f491f6ee8db60e60e3e1cab256cda20de9be81202aa0"} Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.489864 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.491122 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.491564 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" event={"ID":"b48ab67f-88cb-4921-b6a3-38b77b534bf9","Type":"ContainerDied","Data":"c935cf67bf3bb60e71d9e8fa32ea33c10f87bb841cf838b50e9a195acf0ce5f3"} Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.491600 4824 scope.go:117] "RemoveContainer" containerID="415de3e8643b808347c13b2c7d4d57542a514312e1a5bd379cbc7b434241d117" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.491775 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.500893 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" event={"ID":"e8aca65b-76cd-45f9-8b20-da2e34a41fc2","Type":"ContainerDied","Data":"8c626b420278544fc13bdb44a0fc27116c48e8b3b2e62a1f886c8aa8d6e59139"} Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.500969 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.510400 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" podStartSLOduration=10.510386 podStartE2EDuration="10.510386s" podCreationTimestamp="2026-01-21 11:25:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:25:20.498636132 +0000 UTC m=+862.791665423" watchObservedRunningTime="2026-01-21 11:25:20.510386 +0000 UTC m=+862.803415291" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.522703 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" podStartSLOduration=11.522682457 podStartE2EDuration="11.522682457s" podCreationTimestamp="2026-01-21 11:25:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:25:20.517763245 +0000 UTC m=+862.810792537" watchObservedRunningTime="2026-01-21 11:25:20.522682457 +0000 UTC m=+862.815711749" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.545770 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.561132 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.607985 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.613792 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.647199 4824 scope.go:117] "RemoveContainer" containerID="aaacc227843f224a20462553c49fdbfc00aeab8ce2765ad6ee4c575c381c5973" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.647597 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.966787 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-htwd6"] Jan 21 11:25:20 crc kubenswrapper[4824]: W0121 11:25:20.976582 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod511f2b6d_a08d_49f8_b393_ab222219d4a7.slice/crio-6ed1172c8819aecfe47f2a831be778c84a6bcec641f7cd2728eaf87d11c3cd23 WatchSource:0}: Error finding container 6ed1172c8819aecfe47f2a831be778c84a6bcec641f7cd2728eaf87d11c3cd23: Status 404 returned error can't find the container with id 6ed1172c8819aecfe47f2a831be778c84a6bcec641f7cd2728eaf87d11c3cd23 Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.128600 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gr9gj"] Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.245158 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-vrpmw"] Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.246373 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.248176 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vrpmw"] Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.254731 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovs-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282078 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-combined-ca-bundle\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovn-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282134 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8652f3f6-1ad7-49d9-9048-827be17a3dd0-config\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282182 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqtdg\" (UniqueName: \"kubernetes.io/projected/8652f3f6-1ad7-49d9-9048-827be17a3dd0-kube-api-access-sqtdg\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282202 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.330983 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqtdg\" (UniqueName: \"kubernetes.io/projected/8652f3f6-1ad7-49d9-9048-827be17a3dd0-kube-api-access-sqtdg\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383662 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovs-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383701 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-combined-ca-bundle\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383720 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovn-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8652f3f6-1ad7-49d9-9048-827be17a3dd0-config\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.384182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovn-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.384392 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovs-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.384852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8652f3f6-1ad7-49d9-9048-827be17a3dd0-config\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.388591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-combined-ca-bundle\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.388610 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.397918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqtdg\" (UniqueName: \"kubernetes.io/projected/8652f3f6-1ad7-49d9-9048-827be17a3dd0-kube-api-access-sqtdg\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.510482 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-htwd6" event={"ID":"511f2b6d-a08d-49f8-b393-ab222219d4a7","Type":"ContainerStarted","Data":"6ed1172c8819aecfe47f2a831be778c84a6bcec641f7cd2728eaf87d11c3cd23"} Jan 21 11:25:21 crc kubenswrapper[4824]: W0121 11:25:21.549042 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb27b3f77_8476_4f3e_a946_f4f26fc66cb5.slice/crio-b01683b60bc1b3ef97826d5ab6262af5f63c4d702d8d2950e38ddae2e977f4b2 WatchSource:0}: Error finding container b01683b60bc1b3ef97826d5ab6262af5f63c4d702d8d2950e38ddae2e977f4b2: Status 404 returned error can't find the container with id b01683b60bc1b3ef97826d5ab6262af5f63c4d702d8d2950e38ddae2e977f4b2 Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.572572 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.065338 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b48ab67f-88cb-4921-b6a3-38b77b534bf9" path="/var/lib/kubelet/pods/b48ab67f-88cb-4921-b6a3-38b77b534bf9/volumes" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.066621 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8aca65b-76cd-45f9-8b20-da2e34a41fc2" path="/var/lib/kubelet/pods/e8aca65b-76cd-45f9-8b20-da2e34a41fc2/volumes" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.334039 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.351143 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.352148 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.360381 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.371383 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.401539 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.401574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.401862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.401969 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlmf6\" (UniqueName: \"kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.504280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.504339 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.504374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.504968 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.505130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.505663 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.505790 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlmf6\" (UniqueName: \"kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.518571 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gr9gj" event={"ID":"b27b3f77-8476-4f3e-a946-f4f26fc66cb5","Type":"ContainerStarted","Data":"b01683b60bc1b3ef97826d5ab6262af5f63c4d702d8d2950e38ddae2e977f4b2"} Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.518721 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="dnsmasq-dns" containerID="cri-o://0ad321acd47c8d7e7d82f491f6ee8db60e60e3e1cab256cda20de9be81202aa0" gracePeriod=10 Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.518978 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlmf6\" (UniqueName: \"kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.670352 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: W0121 11:25:22.721794 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c0b5876_da04_49e6_8946_4fc6060a64ec.slice/crio-a9120e7f79937c25195d3b830feb6ca41ab891b48425d59e1a0ca2c0018a23cd WatchSource:0}: Error finding container a9120e7f79937c25195d3b830feb6ca41ab891b48425d59e1a0ca2c0018a23cd: Status 404 returned error can't find the container with id a9120e7f79937c25195d3b830feb6ca41ab891b48425d59e1a0ca2c0018a23cd Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.842734 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.844084 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.845724 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.845932 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.846177 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-rn62s" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.847204 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.855280 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911730 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-config\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911788 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911869 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911906 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxnpp\" (UniqueName: \"kubernetes.io/projected/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-kube-api-access-lxnpp\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-config\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014735 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxnpp\" (UniqueName: \"kubernetes.io/projected/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-kube-api-access-lxnpp\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014978 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014992 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.015588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.015663 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-config\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.015886 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.017309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.019371 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.019609 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.023057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.029678 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxnpp\" (UniqueName: \"kubernetes.io/projected/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-kube-api-access-lxnpp\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.047167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.168132 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.525218 4824 generic.go:334] "Generic (PLEG): container finished" podID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerID="0ad321acd47c8d7e7d82f491f6ee8db60e60e3e1cab256cda20de9be81202aa0" exitCode=0 Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.525290 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" event={"ID":"f67dfc52-dc0e-404e-af22-75c6a3d10721","Type":"ContainerDied","Data":"0ad321acd47c8d7e7d82f491f6ee8db60e60e3e1cab256cda20de9be81202aa0"} Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.526128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5c0b5876-da04-49e6-8946-4fc6060a64ec","Type":"ContainerStarted","Data":"a9120e7f79937c25195d3b830feb6ca41ab891b48425d59e1a0ca2c0018a23cd"} Jan 21 11:25:24 crc kubenswrapper[4824]: I0121 11:25:24.952416 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.044268 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc\") pod \"f67dfc52-dc0e-404e-af22-75c6a3d10721\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.044320 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config\") pod \"f67dfc52-dc0e-404e-af22-75c6a3d10721\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.077781 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vrpmw"] Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.091583 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:25 crc kubenswrapper[4824]: W0121 11:25:25.093812 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8652f3f6_1ad7_49d9_9048_827be17a3dd0.slice/crio-352bf16a65066bf3212e9a4c7cf5a4cc64f59beec5baff4e34e6c30d19cd91e0 WatchSource:0}: Error finding container 352bf16a65066bf3212e9a4c7cf5a4cc64f59beec5baff4e34e6c30d19cd91e0: Status 404 returned error can't find the container with id 352bf16a65066bf3212e9a4c7cf5a4cc64f59beec5baff4e34e6c30d19cd91e0 Jan 21 11:25:25 crc kubenswrapper[4824]: W0121 11:25:25.100056 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56a2624d_7ab8_43ea_967e_e4112c693148.slice/crio-bbfbc490e1ea828e4c1b756158c30f088c089cc696efd16dfed5b06bd0756946 WatchSource:0}: Error finding container bbfbc490e1ea828e4c1b756158c30f088c089cc696efd16dfed5b06bd0756946: Status 404 returned error can't find the container with id bbfbc490e1ea828e4c1b756158c30f088c089cc696efd16dfed5b06bd0756946 Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.120439 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.140927 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config" (OuterVolumeSpecName: "config") pod "f67dfc52-dc0e-404e-af22-75c6a3d10721" (UID: "f67dfc52-dc0e-404e-af22-75c6a3d10721"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.145999 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msx57\" (UniqueName: \"kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57\") pod \"f67dfc52-dc0e-404e-af22-75c6a3d10721\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.146508 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.149627 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57" (OuterVolumeSpecName: "kube-api-access-msx57") pod "f67dfc52-dc0e-404e-af22-75c6a3d10721" (UID: "f67dfc52-dc0e-404e-af22-75c6a3d10721"). InnerVolumeSpecName "kube-api-access-msx57". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.174614 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f67dfc52-dc0e-404e-af22-75c6a3d10721" (UID: "f67dfc52-dc0e-404e-af22-75c6a3d10721"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.247406 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msx57\" (UniqueName: \"kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.247433 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.497121 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.544717 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dba0050b-8a73-4355-a1e0-7c9a03557ead","Type":"ContainerStarted","Data":"15764a7d669d69b32030fea1055b01aba131468d39177b7c5dbb1e56a61e22a7"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.545886 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.547864 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2e3a943c-ee06-4e54-a984-34ecb1320f89","Type":"ContainerStarted","Data":"6e0a2a51cb10bfe8bc65ec0123c4fa76489cf9873a673344ebc1fbf8065155d5"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.548355 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.549293 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef","Type":"ContainerStarted","Data":"4ffdf56ad46938e9ac8326bf83cd4631d601d678ad6cb95715b285cb0cfb9687"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.550798 4824 generic.go:334] "Generic (PLEG): container finished" podID="56a2624d-7ab8-43ea-967e-e4112c693148" containerID="7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93" exitCode=0 Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.550843 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" event={"ID":"56a2624d-7ab8-43ea-967e-e4112c693148","Type":"ContainerDied","Data":"7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.550859 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" event={"ID":"56a2624d-7ab8-43ea-967e-e4112c693148","Type":"ContainerStarted","Data":"bbfbc490e1ea828e4c1b756158c30f088c089cc696efd16dfed5b06bd0756946"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.557848 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346","Type":"ContainerStarted","Data":"17209275dbf42e0eb684762f5bbf9fbe138f725455b92becb5b33da1dce19d39"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.561933 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" event={"ID":"f67dfc52-dc0e-404e-af22-75c6a3d10721","Type":"ContainerDied","Data":"4c0d7f6c5bc10bc5e3cf172c7b361f178beb8f1decb24246db808a7099969563"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.562006 4824 scope.go:117] "RemoveContainer" containerID="0ad321acd47c8d7e7d82f491f6ee8db60e60e3e1cab256cda20de9be81202aa0" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.562084 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.563104 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=6.286770926 podStartE2EDuration="11.563090052s" podCreationTimestamp="2026-01-21 11:25:14 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.272123382 +0000 UTC m=+861.565152674" lastFinishedPulling="2026-01-21 11:25:24.548442509 +0000 UTC m=+866.841471800" observedRunningTime="2026-01-21 11:25:25.559134817 +0000 UTC m=+867.852164108" watchObservedRunningTime="2026-01-21 11:25:25.563090052 +0000 UTC m=+867.856119344" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.566411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vrpmw" event={"ID":"8652f3f6-1ad7-49d9-9048-827be17a3dd0","Type":"ContainerStarted","Data":"352bf16a65066bf3212e9a4c7cf5a4cc64f59beec5baff4e34e6c30d19cd91e0"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.575099 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=4.11025354 podStartE2EDuration="9.575084411s" podCreationTimestamp="2026-01-21 11:25:16 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.157110037 +0000 UTC m=+861.450139330" lastFinishedPulling="2026-01-21 11:25:24.621940909 +0000 UTC m=+866.914970201" observedRunningTime="2026-01-21 11:25:25.569416167 +0000 UTC m=+867.862445459" watchObservedRunningTime="2026-01-21 11:25:25.575084411 +0000 UTC m=+867.868113703" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.575773 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"218aca21-4c37-4862-b891-fc228baa5348","Type":"ContainerStarted","Data":"0cdcef9637307812b896082a0b7c8e830ed364eb95d323dde4a524f05b45fc8d"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.695398 4824 scope.go:117] "RemoveContainer" containerID="2fabf442546067371acccd858ab70f06b5ad7ea6a02d270ec355916d978e1384" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.712729 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.717139 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.061518 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" path="/var/lib/kubelet/pods/f67dfc52-dc0e-404e-af22-75c6a3d10721/volumes" Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.582396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerStarted","Data":"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487"} Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.585666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" event={"ID":"56a2624d-7ab8-43ea-967e-e4112c693148","Type":"ContainerStarted","Data":"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa"} Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.586453 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.588682 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerStarted","Data":"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17"} Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.614747 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" podStartSLOduration=4.614732308 podStartE2EDuration="4.614732308s" podCreationTimestamp="2026-01-21 11:25:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:25:26.613557274 +0000 UTC m=+868.906586565" watchObservedRunningTime="2026-01-21 11:25:26.614732308 +0000 UTC m=+868.907761600" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.600180 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5c0b5876-da04-49e6-8946-4fc6060a64ec","Type":"ContainerStarted","Data":"c5bc540e824f8315de4aaa98521390c3084ae1ca6f114bf24ca648548b63b037"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.600221 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5c0b5876-da04-49e6-8946-4fc6060a64ec","Type":"ContainerStarted","Data":"5a4b275ec2f8e3f778381b1131554cd298f52f4fc6b173ea49a8770ea4a60915"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.601829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef","Type":"ContainerStarted","Data":"db803036a9410506d79d5627d9543774fc019eb3ab50fab160d3258604367f1d"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.601869 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef","Type":"ContainerStarted","Data":"719b4c4c84cd5cc9db2a9703e4611c3374447f18dbbb7225f53d8bbfa4a32c07"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.603196 4824 generic.go:334] "Generic (PLEG): container finished" podID="7cd8d62b-ab74-4a1b-95f5-7253ed6c0346" containerID="17209275dbf42e0eb684762f5bbf9fbe138f725455b92becb5b33da1dce19d39" exitCode=0 Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.603253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346","Type":"ContainerDied","Data":"17209275dbf42e0eb684762f5bbf9fbe138f725455b92becb5b33da1dce19d39"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.604564 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vrpmw" event={"ID":"8652f3f6-1ad7-49d9-9048-827be17a3dd0","Type":"ContainerStarted","Data":"aae178d791168024804772a47be73fd13944340220d55c19a31acecf8d0e4018"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.605910 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gr9gj" event={"ID":"b27b3f77-8476-4f3e-a946-f4f26fc66cb5","Type":"ContainerStarted","Data":"f2c83cbd5821aafcf83f1956f57760cd86d073accfd9a9bee311534424f7f71f"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.606970 4824 generic.go:334] "Generic (PLEG): container finished" podID="218aca21-4c37-4862-b891-fc228baa5348" containerID="0cdcef9637307812b896082a0b7c8e830ed364eb95d323dde4a524f05b45fc8d" exitCode=0 Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.607024 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"218aca21-4c37-4862-b891-fc228baa5348","Type":"ContainerDied","Data":"0cdcef9637307812b896082a0b7c8e830ed364eb95d323dde4a524f05b45fc8d"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.610422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-htwd6" event={"ID":"511f2b6d-a08d-49f8-b393-ab222219d4a7","Type":"ContainerStarted","Data":"fbe475794bcef10bb1b0130b5c614e06187deb33b9d1cd49cfa882f3d3c6711d"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.610453 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-htwd6" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.617509 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.078714048 podStartE2EDuration="9.617492879s" podCreationTimestamp="2026-01-21 11:25:19 +0000 UTC" firstStartedPulling="2026-01-21 11:25:22.723951597 +0000 UTC m=+865.016980889" lastFinishedPulling="2026-01-21 11:25:28.262730428 +0000 UTC m=+870.555759720" observedRunningTime="2026-01-21 11:25:28.615473622 +0000 UTC m=+870.908502914" watchObservedRunningTime="2026-01-21 11:25:28.617492879 +0000 UTC m=+870.910522170" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.672394 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-vrpmw" podStartSLOduration=4.489699936 podStartE2EDuration="7.672376763s" podCreationTimestamp="2026-01-21 11:25:21 +0000 UTC" firstStartedPulling="2026-01-21 11:25:25.097757642 +0000 UTC m=+867.390786934" lastFinishedPulling="2026-01-21 11:25:28.280434469 +0000 UTC m=+870.573463761" observedRunningTime="2026-01-21 11:25:28.670622175 +0000 UTC m=+870.963651477" watchObservedRunningTime="2026-01-21 11:25:28.672376763 +0000 UTC m=+870.965406055" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.694179 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.554566417 podStartE2EDuration="7.694162969s" podCreationTimestamp="2026-01-21 11:25:21 +0000 UTC" firstStartedPulling="2026-01-21 11:25:25.128670532 +0000 UTC m=+867.421699824" lastFinishedPulling="2026-01-21 11:25:28.268267085 +0000 UTC m=+870.561296376" observedRunningTime="2026-01-21 11:25:28.688630249 +0000 UTC m=+870.981659541" watchObservedRunningTime="2026-01-21 11:25:28.694162969 +0000 UTC m=+870.987192261" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.711561 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-htwd6" podStartSLOduration=2.43187333 podStartE2EDuration="9.711533842s" podCreationTimestamp="2026-01-21 11:25:19 +0000 UTC" firstStartedPulling="2026-01-21 11:25:20.980756004 +0000 UTC m=+863.273785297" lastFinishedPulling="2026-01-21 11:25:28.260416527 +0000 UTC m=+870.553445809" observedRunningTime="2026-01-21 11:25:28.706154732 +0000 UTC m=+870.999184025" watchObservedRunningTime="2026-01-21 11:25:28.711533842 +0000 UTC m=+871.004563134" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.954804 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.972750 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:28 crc kubenswrapper[4824]: E0121 11:25:28.973058 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="init" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.973075 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="init" Jan 21 11:25:28 crc kubenswrapper[4824]: E0121 11:25:28.973088 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="dnsmasq-dns" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.973094 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="dnsmasq-dns" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.973243 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="dnsmasq-dns" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.973923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.976448 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.990986 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.110405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.110481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.110527 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg5nl\" (UniqueName: \"kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.110635 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.110708 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.168923 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212020 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212182 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212220 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg5nl\" (UniqueName: \"kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212868 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.213172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.213310 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.213353 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.226236 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg5nl\" (UniqueName: \"kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.286492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.594527 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.616971 4824 generic.go:334] "Generic (PLEG): container finished" podID="b27b3f77-8476-4f3e-a946-f4f26fc66cb5" containerID="f2c83cbd5821aafcf83f1956f57760cd86d073accfd9a9bee311534424f7f71f" exitCode=0 Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.617001 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gr9gj" event={"ID":"b27b3f77-8476-4f3e-a946-f4f26fc66cb5","Type":"ContainerDied","Data":"f2c83cbd5821aafcf83f1956f57760cd86d073accfd9a9bee311534424f7f71f"} Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.619603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"218aca21-4c37-4862-b891-fc228baa5348","Type":"ContainerStarted","Data":"b75f9993c8911f138ee27d672fe969e86c04facee2fd173c5eb3fe3d2c104eb2"} Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.621883 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346","Type":"ContainerStarted","Data":"91180231ce54266600bab28049663d812ad89e282ad5a82636c446ff381b8e4c"} Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.622335 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="dnsmasq-dns" containerID="cri-o://9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa" gracePeriod=10 Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.647780 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.657802 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.668527 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=12.364585947 podStartE2EDuration="17.668513293s" podCreationTimestamp="2026-01-21 11:25:12 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.272909113 +0000 UTC m=+861.565938405" lastFinishedPulling="2026-01-21 11:25:24.576836459 +0000 UTC m=+866.869865751" observedRunningTime="2026-01-21 11:25:29.664151141 +0000 UTC m=+871.957180433" watchObservedRunningTime="2026-01-21 11:25:29.668513293 +0000 UTC m=+871.961542586" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.697529 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=13.346555629 podStartE2EDuration="18.697515361s" podCreationTimestamp="2026-01-21 11:25:11 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.273342219 +0000 UTC m=+861.566371511" lastFinishedPulling="2026-01-21 11:25:24.62430195 +0000 UTC m=+866.917331243" observedRunningTime="2026-01-21 11:25:29.694439844 +0000 UTC m=+871.987469136" watchObservedRunningTime="2026-01-21 11:25:29.697515361 +0000 UTC m=+871.990544653" Jan 21 11:25:29 crc kubenswrapper[4824]: W0121 11:25:29.712046 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod292b341f_0f9e_411e_9f85_47bde4dcb2c4.slice/crio-c0704ee36eacd59c7b38561ac33e2d4956e244055bb28267047fee7b43cd9db0 WatchSource:0}: Error finding container c0704ee36eacd59c7b38561ac33e2d4956e244055bb28267047fee7b43cd9db0: Status 404 returned error can't find the container with id c0704ee36eacd59c7b38561ac33e2d4956e244055bb28267047fee7b43cd9db0 Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.034674 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.127085 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config\") pod \"56a2624d-7ab8-43ea-967e-e4112c693148\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.127137 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb\") pod \"56a2624d-7ab8-43ea-967e-e4112c693148\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.127192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlmf6\" (UniqueName: \"kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6\") pod \"56a2624d-7ab8-43ea-967e-e4112c693148\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.127231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc\") pod \"56a2624d-7ab8-43ea-967e-e4112c693148\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.131367 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6" (OuterVolumeSpecName: "kube-api-access-rlmf6") pod "56a2624d-7ab8-43ea-967e-e4112c693148" (UID: "56a2624d-7ab8-43ea-967e-e4112c693148"). InnerVolumeSpecName "kube-api-access-rlmf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.154608 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "56a2624d-7ab8-43ea-967e-e4112c693148" (UID: "56a2624d-7ab8-43ea-967e-e4112c693148"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.155528 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config" (OuterVolumeSpecName: "config") pod "56a2624d-7ab8-43ea-967e-e4112c693148" (UID: "56a2624d-7ab8-43ea-967e-e4112c693148"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.157172 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "56a2624d-7ab8-43ea-967e-e4112c693148" (UID: "56a2624d-7ab8-43ea-967e-e4112c693148"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.229027 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.229061 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.229075 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlmf6\" (UniqueName: \"kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.229084 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.627701 4824 generic.go:334] "Generic (PLEG): container finished" podID="56a2624d-7ab8-43ea-967e-e4112c693148" containerID="9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa" exitCode=0 Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.627773 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.627792 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" event={"ID":"56a2624d-7ab8-43ea-967e-e4112c693148","Type":"ContainerDied","Data":"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.628206 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" event={"ID":"56a2624d-7ab8-43ea-967e-e4112c693148","Type":"ContainerDied","Data":"bbfbc490e1ea828e4c1b756158c30f088c089cc696efd16dfed5b06bd0756946"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.628224 4824 scope.go:117] "RemoveContainer" containerID="9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.629311 4824 generic.go:334] "Generic (PLEG): container finished" podID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerID="b9f66e73f6529373d74d9e19ba52303b84fabf65da4786d896c78db5e897096f" exitCode=0 Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.629359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" event={"ID":"292b341f-0f9e-411e-9f85-47bde4dcb2c4","Type":"ContainerDied","Data":"b9f66e73f6529373d74d9e19ba52303b84fabf65da4786d896c78db5e897096f"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.629382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" event={"ID":"292b341f-0f9e-411e-9f85-47bde4dcb2c4","Type":"ContainerStarted","Data":"c0704ee36eacd59c7b38561ac33e2d4956e244055bb28267047fee7b43cd9db0"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.632079 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gr9gj" event={"ID":"b27b3f77-8476-4f3e-a946-f4f26fc66cb5","Type":"ContainerStarted","Data":"377f978847f48f8fced112de3ef20125036e82a1968c553b3962be0500956157"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.632132 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gr9gj" event={"ID":"b27b3f77-8476-4f3e-a946-f4f26fc66cb5","Type":"ContainerStarted","Data":"2aa2e9ec59322b23149f826a199fa37ef4c7eedab4b46616599082f2a3764f7c"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.632313 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.632451 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.644704 4824 scope.go:117] "RemoveContainer" containerID="7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.647809 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.648146 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-gr9gj" podStartSLOduration=4.955803298 podStartE2EDuration="11.648136056s" podCreationTimestamp="2026-01-21 11:25:19 +0000 UTC" firstStartedPulling="2026-01-21 11:25:21.550696857 +0000 UTC m=+863.843726148" lastFinishedPulling="2026-01-21 11:25:28.243029613 +0000 UTC m=+870.536058906" observedRunningTime="2026-01-21 11:25:30.647982939 +0000 UTC m=+872.941012230" watchObservedRunningTime="2026-01-21 11:25:30.648136056 +0000 UTC m=+872.941165348" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.718697 4824 scope.go:117] "RemoveContainer" containerID="9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa" Jan 21 11:25:30 crc kubenswrapper[4824]: E0121 11:25:30.719826 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa\": container with ID starting with 9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa not found: ID does not exist" containerID="9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.719885 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa"} err="failed to get container status \"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa\": rpc error: code = NotFound desc = could not find container \"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa\": container with ID starting with 9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa not found: ID does not exist" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.719907 4824 scope.go:117] "RemoveContainer" containerID="7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93" Jan 21 11:25:30 crc kubenswrapper[4824]: E0121 11:25:30.720308 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93\": container with ID starting with 7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93 not found: ID does not exist" containerID="7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.720341 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93"} err="failed to get container status \"7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93\": rpc error: code = NotFound desc = could not find container \"7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93\": container with ID starting with 7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93 not found: ID does not exist" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.761910 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.767278 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:31 crc kubenswrapper[4824]: I0121 11:25:31.639286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" event={"ID":"292b341f-0f9e-411e-9f85-47bde4dcb2c4","Type":"ContainerStarted","Data":"095e6c8fd89e1deec2822bbad9ce47d3bc5541769b4bdf66512d4544bcb46405"} Jan 21 11:25:31 crc kubenswrapper[4824]: I0121 11:25:31.656346 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" podStartSLOduration=3.656333091 podStartE2EDuration="3.656333091s" podCreationTimestamp="2026-01-21 11:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:25:31.651936754 +0000 UTC m=+873.944966046" watchObservedRunningTime="2026-01-21 11:25:31.656333091 +0000 UTC m=+873.949362383" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.056409 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" path="/var/lib/kubelet/pods/56a2624d-7ab8-43ea-967e-e4112c693148/volumes" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.195011 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.195418 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.644091 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.676333 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.975784 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.975822 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Jan 21 11:25:33 crc kubenswrapper[4824]: I0121 11:25:33.191771 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Jan 21 11:25:33 crc kubenswrapper[4824]: I0121 11:25:33.193120 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:33 crc kubenswrapper[4824]: I0121 11:25:33.695328 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.269223 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jbk7q"] Jan 21 11:25:34 crc kubenswrapper[4824]: E0121 11:25:34.269501 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="init" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.269516 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="init" Jan 21 11:25:34 crc kubenswrapper[4824]: E0121 11:25:34.269546 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="dnsmasq-dns" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.269554 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="dnsmasq-dns" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.269705 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="dnsmasq-dns" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.270128 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.275966 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jbk7q"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.319676 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-383b-account-create-update-w99bv"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.320508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.322065 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.324559 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.324595 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.327902 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-383b-account-create-update-w99bv"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.373502 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.382989 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.383042 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcw6h\" (UniqueName: \"kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.484786 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.484867 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcw6h\" (UniqueName: \"kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.485035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.485228 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2qxk\" (UniqueName: \"kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.485401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.499666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcw6h\" (UniqueName: \"kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.521130 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-q7wdg"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.522211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.527058 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-q7wdg"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.531835 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a0fd-account-create-update-zxjvg"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.533174 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.534372 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.541032 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a0fd-account-create-update-zxjvg"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.583988 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.586232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.586334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2qxk\" (UniqueName: \"kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.587127 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.599554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2qxk\" (UniqueName: \"kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.636430 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.688376 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.690159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z72x\" (UniqueName: \"kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.690244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.690270 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvg8j\" (UniqueName: \"kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.738474 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.791943 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z72x\" (UniqueName: \"kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.792041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.792059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvg8j\" (UniqueName: \"kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.792135 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.794491 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.794501 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.807274 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvg8j\" (UniqueName: \"kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.807505 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z72x\" (UniqueName: \"kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.848798 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.857212 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.947425 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jbk7q"] Jan 21 11:25:34 crc kubenswrapper[4824]: W0121 11:25:34.948534 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c01cf79_8a28_4229_9257_90124e66d3cc.slice/crio-1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8 WatchSource:0}: Error finding container 1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8: Status 404 returned error can't find the container with id 1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.035710 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-383b-account-create-update-w99bv"] Jan 21 11:25:35 crc kubenswrapper[4824]: W0121 11:25:35.042775 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc853ab0_e92a_4ada_84fd_b59ac2af573d.slice/crio-43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5 WatchSource:0}: Error finding container 43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5: Status 404 returned error can't find the container with id 43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.222634 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a0fd-account-create-update-zxjvg"] Jan 21 11:25:35 crc kubenswrapper[4824]: W0121 11:25:35.230574 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0ca5b98_4b63_4b4e_a80f_d52977ae3d01.slice/crio-a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66 WatchSource:0}: Error finding container a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66: Status 404 returned error can't find the container with id a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.276939 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-q7wdg"] Jan 21 11:25:35 crc kubenswrapper[4824]: W0121 11:25:35.280745 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod885db96c_779f_47c5_a383_1821cb53977c.slice/crio-a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f WatchSource:0}: Error finding container a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f: Status 404 returned error can't find the container with id a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.660715 4824 generic.go:334] "Generic (PLEG): container finished" podID="bc853ab0-e92a-4ada-84fd-b59ac2af573d" containerID="0811ac195531b85d7ce2dbbbbfe5e4b4eec9c6e973dbfd91e9e942aecc0be98b" exitCode=0 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.660769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-383b-account-create-update-w99bv" event={"ID":"bc853ab0-e92a-4ada-84fd-b59ac2af573d","Type":"ContainerDied","Data":"0811ac195531b85d7ce2dbbbbfe5e4b4eec9c6e973dbfd91e9e942aecc0be98b"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.660791 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-383b-account-create-update-w99bv" event={"ID":"bc853ab0-e92a-4ada-84fd-b59ac2af573d","Type":"ContainerStarted","Data":"43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.662260 4824 generic.go:334] "Generic (PLEG): container finished" podID="a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" containerID="be822998cfd0d1fe00b165b7bdf2ce2c669d12e3704320cfa573970aa648494d" exitCode=0 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.662296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0fd-account-create-update-zxjvg" event={"ID":"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01","Type":"ContainerDied","Data":"be822998cfd0d1fe00b165b7bdf2ce2c669d12e3704320cfa573970aa648494d"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.662312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0fd-account-create-update-zxjvg" event={"ID":"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01","Type":"ContainerStarted","Data":"a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.665872 4824 generic.go:334] "Generic (PLEG): container finished" podID="7c01cf79-8a28-4229-9257-90124e66d3cc" containerID="e85df67ff5e42c058f160ae767051c544cbf1f8e15561d3884068b5517d68c3a" exitCode=0 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.665914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jbk7q" event={"ID":"7c01cf79-8a28-4229-9257-90124e66d3cc","Type":"ContainerDied","Data":"e85df67ff5e42c058f160ae767051c544cbf1f8e15561d3884068b5517d68c3a"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.665930 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jbk7q" event={"ID":"7c01cf79-8a28-4229-9257-90124e66d3cc","Type":"ContainerStarted","Data":"1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.667174 4824 generic.go:334] "Generic (PLEG): container finished" podID="885db96c-779f-47c5-a383-1821cb53977c" containerID="7d54dd5634955be1b99e4810715039653beb466487ffbc563c911444570d3daa" exitCode=0 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.669779 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q7wdg" event={"ID":"885db96c-779f-47c5-a383-1821cb53977c","Type":"ContainerDied","Data":"7d54dd5634955be1b99e4810715039653beb466487ffbc563c911444570d3daa"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.669820 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q7wdg" event={"ID":"885db96c-779f-47c5-a383-1821cb53977c","Type":"ContainerStarted","Data":"a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.676693 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.851398 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.852723 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.855005 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.855176 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.855304 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-qc9wz" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.858047 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.874570 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.012513 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.012590 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-config\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.012676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-scripts\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.012832 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.012916 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27w66\" (UniqueName: \"kubernetes.io/projected/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-kube-api-access-27w66\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.013064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.013084 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.113851 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27w66\" (UniqueName: \"kubernetes.io/projected/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-kube-api-access-27w66\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.113942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.113974 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.114002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.114054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-config\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.114068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-scripts\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.114117 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.114523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.115975 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-config\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.116489 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-scripts\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.119057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.121442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.127800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27w66\" (UniqueName: \"kubernetes.io/projected/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-kube-api-access-27w66\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.134897 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.164041 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.328429 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.328823 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="dnsmasq-dns" containerID="cri-o://095e6c8fd89e1deec2822bbad9ce47d3bc5541769b4bdf66512d4544bcb46405" gracePeriod=10 Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.333675 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.359085 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.374381 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.374468 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.432785 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.526063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.526131 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9jhb\" (UniqueName: \"kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.526242 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.526279 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.526319 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.606218 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.627658 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.627772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.627808 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9jhb\" (UniqueName: \"kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.627857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.627886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.628601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.629121 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.629183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.629536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.641860 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9jhb\" (UniqueName: \"kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.678333 4824 generic.go:334] "Generic (PLEG): container finished" podID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerID="095e6c8fd89e1deec2822bbad9ce47d3bc5541769b4bdf66512d4544bcb46405" exitCode=0 Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.678409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" event={"ID":"292b341f-0f9e-411e-9f85-47bde4dcb2c4","Type":"ContainerDied","Data":"095e6c8fd89e1deec2822bbad9ce47d3bc5541769b4bdf66512d4544bcb46405"} Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.679432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b9c25b1-d0b9-430e-88cf-0aeaf6785461","Type":"ContainerStarted","Data":"9247690bf5598a07e7dc894f7f927cbb898894b6d00fd4477369b84588389a75"} Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.694611 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.987202 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.135380 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts\") pod \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.135663 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2qxk\" (UniqueName: \"kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk\") pod \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.136088 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc853ab0-e92a-4ada-84fd-b59ac2af573d" (UID: "bc853ab0-e92a-4ada-84fd-b59ac2af573d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.136212 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.140612 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk" (OuterVolumeSpecName: "kube-api-access-t2qxk") pod "bc853ab0-e92a-4ada-84fd-b59ac2af573d" (UID: "bc853ab0-e92a-4ada-84fd-b59ac2af573d"). InnerVolumeSpecName "kube-api-access-t2qxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.167709 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.181205 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.187931 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.237402 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2qxk\" (UniqueName: \"kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.268031 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:25:37 crc kubenswrapper[4824]: W0121 11:25:37.269498 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode94778b1_d36a_47e9_b201_9b7cd1cdb640.slice/crio-48fb0631b23867cac173ec7a9df005de8edd1cb0df987f8dceb763f2434dd5b7 WatchSource:0}: Error finding container 48fb0631b23867cac173ec7a9df005de8edd1cb0df987f8dceb763f2434dd5b7: Status 404 returned error can't find the container with id 48fb0631b23867cac173ec7a9df005de8edd1cb0df987f8dceb763f2434dd5b7 Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvg8j\" (UniqueName: \"kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j\") pod \"885db96c-779f-47c5-a383-1821cb53977c\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338358 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts\") pod \"885db96c-779f-47c5-a383-1821cb53977c\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338393 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts\") pod \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338429 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts\") pod \"7c01cf79-8a28-4229-9257-90124e66d3cc\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338503 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2z72x\" (UniqueName: \"kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x\") pod \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338530 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcw6h\" (UniqueName: \"kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h\") pod \"7c01cf79-8a28-4229-9257-90124e66d3cc\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.339140 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" (UID: "a0ca5b98-4b63-4b4e-a80f-d52977ae3d01"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.339159 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7c01cf79-8a28-4229-9257-90124e66d3cc" (UID: "7c01cf79-8a28-4229-9257-90124e66d3cc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.339195 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "885db96c-779f-47c5-a383-1821cb53977c" (UID: "885db96c-779f-47c5-a383-1821cb53977c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.341099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x" (OuterVolumeSpecName: "kube-api-access-2z72x") pod "a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" (UID: "a0ca5b98-4b63-4b4e-a80f-d52977ae3d01"). InnerVolumeSpecName "kube-api-access-2z72x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.342394 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h" (OuterVolumeSpecName: "kube-api-access-hcw6h") pod "7c01cf79-8a28-4229-9257-90124e66d3cc" (UID: "7c01cf79-8a28-4229-9257-90124e66d3cc"). InnerVolumeSpecName "kube-api-access-hcw6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.343359 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j" (OuterVolumeSpecName: "kube-api-access-cvg8j") pod "885db96c-779f-47c5-a383-1821cb53977c" (UID: "885db96c-779f-47c5-a383-1821cb53977c"). InnerVolumeSpecName "kube-api-access-cvg8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440740 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvg8j\" (UniqueName: \"kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440770 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440779 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440788 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440801 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2z72x\" (UniqueName: \"kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440808 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcw6h\" (UniqueName: \"kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.510585 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.510879 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="885db96c-779f-47c5-a383-1821cb53977c" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.510896 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="885db96c-779f-47c5-a383-1821cb53977c" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.510918 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c01cf79-8a28-4229-9257-90124e66d3cc" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.510924 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c01cf79-8a28-4229-9257-90124e66d3cc" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.510952 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.510972 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.510987 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc853ab0-e92a-4ada-84fd-b59ac2af573d" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.510992 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc853ab0-e92a-4ada-84fd-b59ac2af573d" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.511124 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c01cf79-8a28-4229-9257-90124e66d3cc" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.511143 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc853ab0-e92a-4ada-84fd-b59ac2af573d" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.511151 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="885db96c-779f-47c5-a383-1821cb53977c" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.511158 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.514744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.516392 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-lk8t8" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.517136 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.517144 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.517218 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.528089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.643139 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-lock\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.643180 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-cache\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.643197 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hwq7\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-kube-api-access-5hwq7\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.643226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.643354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.685953 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.685950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0fd-account-create-update-zxjvg" event={"ID":"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01","Type":"ContainerDied","Data":"a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66"} Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.686262 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.687351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jbk7q" event={"ID":"7c01cf79-8a28-4229-9257-90124e66d3cc","Type":"ContainerDied","Data":"1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8"} Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.687383 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.687385 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.688486 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q7wdg" event={"ID":"885db96c-779f-47c5-a383-1821cb53977c","Type":"ContainerDied","Data":"a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f"} Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.688514 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.688556 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.690465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qzfrf" event={"ID":"e94778b1-d36a-47e9-b201-9b7cd1cdb640","Type":"ContainerStarted","Data":"48fb0631b23867cac173ec7a9df005de8edd1cb0df987f8dceb763f2434dd5b7"} Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.691612 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-383b-account-create-update-w99bv" event={"ID":"bc853ab0-e92a-4ada-84fd-b59ac2af573d","Type":"ContainerDied","Data":"43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5"} Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.691646 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.691697 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745171 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.745352 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.745376 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.745419 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift podName:f980e493-ca74-47f1-8f6f-aca9c9011bf5 nodeName:}" failed. No retries permitted until 2026-01-21 11:25:38.24540553 +0000 UTC m=+880.538434821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift") pod "swift-storage-0" (UID: "f980e493-ca74-47f1-8f6f-aca9c9011bf5") : configmap "swift-ring-files" not found Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-lock\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-cache\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745532 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hwq7\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-kube-api-access-5hwq7\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745815 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745864 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-lock\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745892 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-cache\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.759598 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hwq7\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-kube-api-access-5hwq7\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.765361 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:38 crc kubenswrapper[4824]: I0121 11:25:38.253101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:38 crc kubenswrapper[4824]: E0121 11:25:38.253294 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 21 11:25:38 crc kubenswrapper[4824]: E0121 11:25:38.253317 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 21 11:25:38 crc kubenswrapper[4824]: E0121 11:25:38.253366 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift podName:f980e493-ca74-47f1-8f6f-aca9c9011bf5 nodeName:}" failed. No retries permitted until 2026-01-21 11:25:39.253351654 +0000 UTC m=+881.546380946 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift") pod "swift-storage-0" (UID: "f980e493-ca74-47f1-8f6f-aca9c9011bf5") : configmap "swift-ring-files" not found Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.266830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:39 crc kubenswrapper[4824]: E0121 11:25:39.267002 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 21 11:25:39 crc kubenswrapper[4824]: E0121 11:25:39.267021 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 21 11:25:39 crc kubenswrapper[4824]: E0121 11:25:39.267063 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift podName:f980e493-ca74-47f1-8f6f-aca9c9011bf5 nodeName:}" failed. No retries permitted until 2026-01-21 11:25:41.267051492 +0000 UTC m=+883.560080784 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift") pod "swift-storage-0" (UID: "f980e493-ca74-47f1-8f6f-aca9c9011bf5") : configmap "swift-ring-files" not found Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.287525 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.106:5353: connect: connection refused" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.775142 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-rlxbb"] Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.776196 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.784807 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rlxbb"] Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.875023 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-0697-account-create-update-9nvnh"] Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.875796 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.875849 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxqbh\" (UniqueName: \"kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.875891 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.877854 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.882320 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0697-account-create-update-9nvnh"] Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.977830 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4wtz\" (UniqueName: \"kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.977877 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.977927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.978013 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxqbh\" (UniqueName: \"kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.978541 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.993273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxqbh\" (UniqueName: \"kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.078973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4wtz\" (UniqueName: \"kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.079017 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.079616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.091871 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.092100 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4wtz\" (UniqueName: \"kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.187135 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.457594 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rlxbb"] Jan 21 11:25:40 crc kubenswrapper[4824]: W0121 11:25:40.460428 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fbc8b12_d965_44e9_822f_b95a5c5e88ab.slice/crio-13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf WatchSource:0}: Error finding container 13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf: Status 404 returned error can't find the container with id 13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.551783 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0697-account-create-update-9nvnh"] Jan 21 11:25:40 crc kubenswrapper[4824]: W0121 11:25:40.554622 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72452c85_3787_4bd5_b605_af5555c1fdaf.slice/crio-87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa WatchSource:0}: Error finding container 87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa: Status 404 returned error can't find the container with id 87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.708419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rlxbb" event={"ID":"6fbc8b12-d965-44e9-822f-b95a5c5e88ab","Type":"ContainerStarted","Data":"13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf"} Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.709409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0697-account-create-update-9nvnh" event={"ID":"72452c85-3787-4bd5-b605-af5555c1fdaf","Type":"ContainerStarted","Data":"87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.296953 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.297086 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.297220 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.297470 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift podName:f980e493-ca74-47f1-8f6f-aca9c9011bf5 nodeName:}" failed. No retries permitted until 2026-01-21 11:25:45.297453385 +0000 UTC m=+887.590482678 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift") pod "swift-storage-0" (UID: "f980e493-ca74-47f1-8f6f-aca9c9011bf5") : configmap "swift-ring-files" not found Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.368072 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.458610 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-w2dc9"] Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.458937 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="init" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.458950 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="init" Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.459011 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="dnsmasq-dns" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.459017 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="dnsmasq-dns" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.459161 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="dnsmasq-dns" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.459619 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.463087 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.463096 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.463130 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.501419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb\") pod \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.501500 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc\") pod \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.501551 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg5nl\" (UniqueName: \"kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl\") pod \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.501673 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config\") pod \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.501722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb\") pod \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.507415 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl" (OuterVolumeSpecName: "kube-api-access-hg5nl") pod "292b341f-0f9e-411e-9f85-47bde4dcb2c4" (UID: "292b341f-0f9e-411e-9f85-47bde4dcb2c4"). InnerVolumeSpecName "kube-api-access-hg5nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.507492 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-w2dc9"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.510677 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-nc6vf"] Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.511080 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-ssxbb ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-ssxbb ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-w2dc9" podUID="d65692e9-ea67-4312-b382-dbe3442c3a32" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.511825 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.514754 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-w2dc9"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.519321 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nc6vf"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.548144 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config" (OuterVolumeSpecName: "config") pod "292b341f-0f9e-411e-9f85-47bde4dcb2c4" (UID: "292b341f-0f9e-411e-9f85-47bde4dcb2c4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.550347 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "292b341f-0f9e-411e-9f85-47bde4dcb2c4" (UID: "292b341f-0f9e-411e-9f85-47bde4dcb2c4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.550555 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "292b341f-0f9e-411e-9f85-47bde4dcb2c4" (UID: "292b341f-0f9e-411e-9f85-47bde4dcb2c4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.554360 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "292b341f-0f9e-411e-9f85-47bde4dcb2c4" (UID: "292b341f-0f9e-411e-9f85-47bde4dcb2c4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603151 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssxbb\" (UniqueName: \"kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603217 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cffv8\" (UniqueName: \"kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603273 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603307 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603322 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603348 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603374 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603427 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603597 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603612 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603621 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603629 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603637 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg5nl\" (UniqueName: \"kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.626043 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-n665r"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.626913 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.629042 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.633267 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-n665r"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssxbb\" (UniqueName: \"kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704589 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cffv8\" (UniqueName: \"kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704614 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704724 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704902 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704951 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m496\" (UniqueName: \"kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705412 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705931 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.706460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.708704 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.709224 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.709791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.709807 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.710622 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.711235 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.718186 4824 generic.go:334] "Generic (PLEG): container finished" podID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerID="a517f23f781efca10654bdff86d5e2efc32071ea09d5af5829095603d962c629" exitCode=0 Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.718242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qzfrf" event={"ID":"e94778b1-d36a-47e9-b201-9b7cd1cdb640","Type":"ContainerDied","Data":"a517f23f781efca10654bdff86d5e2efc32071ea09d5af5829095603d962c629"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.719328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssxbb\" (UniqueName: \"kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.719884 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cffv8\" (UniqueName: \"kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.720482 4824 generic.go:334] "Generic (PLEG): container finished" podID="72452c85-3787-4bd5-b605-af5555c1fdaf" containerID="aece7515cad2f167dd4b5a5e58d5aa5c0df426c12c8bbc743e03dad13feaedf7" exitCode=0 Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.720513 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0697-account-create-update-9nvnh" event={"ID":"72452c85-3787-4bd5-b605-af5555c1fdaf","Type":"ContainerDied","Data":"aece7515cad2f167dd4b5a5e58d5aa5c0df426c12c8bbc743e03dad13feaedf7"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.723781 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" event={"ID":"292b341f-0f9e-411e-9f85-47bde4dcb2c4","Type":"ContainerDied","Data":"c0704ee36eacd59c7b38561ac33e2d4956e244055bb28267047fee7b43cd9db0"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.723820 4824 scope.go:117] "RemoveContainer" containerID="095e6c8fd89e1deec2822bbad9ce47d3bc5541769b4bdf66512d4544bcb46405" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.723940 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.726227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b9c25b1-d0b9-430e-88cf-0aeaf6785461","Type":"ContainerStarted","Data":"301fce5be639a4edea0791738f7bedb2f7586b53e5f9101d69276726401d0a51"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.726258 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b9c25b1-d0b9-430e-88cf-0aeaf6785461","Type":"ContainerStarted","Data":"6dceef68019a4aed59819b21a85c3314c7f535000cc74c2484dafc2002f1d644"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.727313 4824 generic.go:334] "Generic (PLEG): container finished" podID="6fbc8b12-d965-44e9-822f-b95a5c5e88ab" containerID="091c53575308b0937e0789468c0edd3eb80a63b653591b4cc1559689a8b4b286" exitCode=0 Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.727463 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.728227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rlxbb" event={"ID":"6fbc8b12-d965-44e9-822f-b95a5c5e88ab","Type":"ContainerDied","Data":"091c53575308b0937e0789468c0edd3eb80a63b653591b4cc1559689a8b4b286"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.741895 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.749903 4824 scope.go:117] "RemoveContainer" containerID="b9f66e73f6529373d74d9e19ba52303b84fabf65da4786d896c78db5e897096f" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.755538 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.106391518 podStartE2EDuration="6.755523996s" podCreationTimestamp="2026-01-21 11:25:35 +0000 UTC" firstStartedPulling="2026-01-21 11:25:36.613513837 +0000 UTC m=+878.906543129" lastFinishedPulling="2026-01-21 11:25:41.262646314 +0000 UTC m=+883.555675607" observedRunningTime="2026-01-21 11:25:41.751742098 +0000 UTC m=+884.044771391" watchObservedRunningTime="2026-01-21 11:25:41.755523996 +0000 UTC m=+884.048553288" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.788726 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.793864 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.806738 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.806893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m496\" (UniqueName: \"kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.807420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.821487 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m496\" (UniqueName: \"kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.838616 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910068 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910326 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssxbb\" (UniqueName: \"kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910461 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910702 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910758 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910785 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts" (OuterVolumeSpecName: "scripts") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910991 4824 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.911004 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.911013 4824 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.912730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.912830 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb" (OuterVolumeSpecName: "kube-api-access-ssxbb") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "kube-api-access-ssxbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.913053 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.914294 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.940281 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n665r" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.012147 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssxbb\" (UniqueName: \"kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.012173 4824 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.012183 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.012191 4824 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.058603 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" path="/var/lib/kubelet/pods/292b341f-0f9e-411e-9f85-47bde4dcb2c4/volumes" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.190551 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nc6vf"] Jan 21 11:25:42 crc kubenswrapper[4824]: W0121 11:25:42.194483 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29473359_3517_4b86_bd5e_80e25706ff27.slice/crio-be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0 WatchSource:0}: Error finding container be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0: Status 404 returned error can't find the container with id be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0 Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.294381 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-n665r"] Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.734917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qzfrf" event={"ID":"e94778b1-d36a-47e9-b201-9b7cd1cdb640","Type":"ContainerStarted","Data":"8a0b04dc02e5d8c871ec744b980efd9b4e195f4024e4ef8187ce38c2f8284265"} Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.735146 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.735778 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nc6vf" event={"ID":"29473359-3517-4b86-bd5e-80e25706ff27","Type":"ContainerStarted","Data":"be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0"} Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.736830 4824 generic.go:334] "Generic (PLEG): container finished" podID="af5c64ab-b4a4-43f5-81c3-416194e40b9a" containerID="26ba4c23450a1625fb62c54cdc290126e5f1371f38bf091a20d00cb1c946e58f" exitCode=0 Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.736868 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n665r" event={"ID":"af5c64ab-b4a4-43f5-81c3-416194e40b9a","Type":"ContainerDied","Data":"26ba4c23450a1625fb62c54cdc290126e5f1371f38bf091a20d00cb1c946e58f"} Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.736886 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n665r" event={"ID":"af5c64ab-b4a4-43f5-81c3-416194e40b9a","Type":"ContainerStarted","Data":"ad7dafc5bf4c4bfe5a803af9f930e427405aeb6acf9bb0bca8e9c155247c42ed"} Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.736895 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.737271 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.756060 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-qzfrf" podStartSLOduration=6.756046675 podStartE2EDuration="6.756046675s" podCreationTimestamp="2026-01-21 11:25:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:25:42.752804905 +0000 UTC m=+885.045834197" watchObservedRunningTime="2026-01-21 11:25:42.756046675 +0000 UTC m=+885.049075967" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.777267 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-w2dc9"] Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.783666 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-w2dc9"] Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.102481 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.159123 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.231043 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxqbh\" (UniqueName: \"kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh\") pod \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.231115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4wtz\" (UniqueName: \"kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz\") pod \"72452c85-3787-4bd5-b605-af5555c1fdaf\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.231159 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts\") pod \"72452c85-3787-4bd5-b605-af5555c1fdaf\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.231237 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts\") pod \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.231699 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "72452c85-3787-4bd5-b605-af5555c1fdaf" (UID: "72452c85-3787-4bd5-b605-af5555c1fdaf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.232121 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6fbc8b12-d965-44e9-822f-b95a5c5e88ab" (UID: "6fbc8b12-d965-44e9-822f-b95a5c5e88ab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.235696 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh" (OuterVolumeSpecName: "kube-api-access-hxqbh") pod "6fbc8b12-d965-44e9-822f-b95a5c5e88ab" (UID: "6fbc8b12-d965-44e9-822f-b95a5c5e88ab"). InnerVolumeSpecName "kube-api-access-hxqbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.235739 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz" (OuterVolumeSpecName: "kube-api-access-g4wtz") pod "72452c85-3787-4bd5-b605-af5555c1fdaf" (UID: "72452c85-3787-4bd5-b605-af5555c1fdaf"). InnerVolumeSpecName "kube-api-access-g4wtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.295854 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:43 crc kubenswrapper[4824]: E0121 11:25:43.296270 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fbc8b12-d965-44e9-822f-b95a5c5e88ab" containerName="mariadb-database-create" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.296299 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fbc8b12-d965-44e9-822f-b95a5c5e88ab" containerName="mariadb-database-create" Jan 21 11:25:43 crc kubenswrapper[4824]: E0121 11:25:43.296319 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72452c85-3787-4bd5-b605-af5555c1fdaf" containerName="mariadb-account-create-update" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.296324 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72452c85-3787-4bd5-b605-af5555c1fdaf" containerName="mariadb-account-create-update" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.296535 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fbc8b12-d965-44e9-822f-b95a5c5e88ab" containerName="mariadb-database-create" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.296547 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="72452c85-3787-4bd5-b605-af5555c1fdaf" containerName="mariadb-account-create-update" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.297817 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.305178 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.333476 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4wtz\" (UniqueName: \"kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.333502 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.334603 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.334614 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxqbh\" (UniqueName: \"kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.435828 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.435999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mmv9\" (UniqueName: \"kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.436040 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.537763 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mmv9\" (UniqueName: \"kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.537831 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.537950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.538324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.538558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.559698 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mmv9\" (UniqueName: \"kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.622781 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.744728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0697-account-create-update-9nvnh" event={"ID":"72452c85-3787-4bd5-b605-af5555c1fdaf","Type":"ContainerDied","Data":"87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa"} Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.744771 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.744745 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.745938 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.746418 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rlxbb" event={"ID":"6fbc8b12-d965-44e9-822f-b95a5c5e88ab","Type":"ContainerDied","Data":"13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf"} Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.746470 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.056237 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d65692e9-ea67-4312-b382-dbe3442c3a32" path="/var/lib/kubelet/pods/d65692e9-ea67-4312-b382-dbe3442c3a32/volumes" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.753828 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n665r" event={"ID":"af5c64ab-b4a4-43f5-81c3-416194e40b9a","Type":"ContainerDied","Data":"ad7dafc5bf4c4bfe5a803af9f930e427405aeb6acf9bb0bca8e9c155247c42ed"} Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.753867 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad7dafc5bf4c4bfe5a803af9f930e427405aeb6acf9bb0bca8e9c155247c42ed" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.868864 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n665r" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.961851 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m496\" (UniqueName: \"kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496\") pod \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.961968 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts\") pod \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.962892 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af5c64ab-b4a4-43f5-81c3-416194e40b9a" (UID: "af5c64ab-b4a4-43f5-81c3-416194e40b9a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.966069 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496" (OuterVolumeSpecName: "kube-api-access-8m496") pod "af5c64ab-b4a4-43f5-81c3-416194e40b9a" (UID: "af5c64ab-b4a4-43f5-81c3-416194e40b9a"). InnerVolumeSpecName "kube-api-access-8m496". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.998476 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-sx9lb"] Jan 21 11:25:44 crc kubenswrapper[4824]: E0121 11:25:44.998796 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af5c64ab-b4a4-43f5-81c3-416194e40b9a" containerName="mariadb-account-create-update" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.998813 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="af5c64ab-b4a4-43f5-81c3-416194e40b9a" containerName="mariadb-account-create-update" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.998981 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="af5c64ab-b4a4-43f5-81c3-416194e40b9a" containerName="mariadb-account-create-update" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.999437 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.001336 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4bzft" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.001536 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.006546 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sx9lb"] Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.063355 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m496\" (UniqueName: \"kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.063381 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.128543 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:45 crc kubenswrapper[4824]: W0121 11:25:45.135154 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb4180b2_5d3e_486d_8197_348d01d2286e.slice/crio-f00978733ebd4e9d3763dd20389763b97d4ebd2a17854c4cfba91b3e800bb66e WatchSource:0}: Error finding container f00978733ebd4e9d3763dd20389763b97d4ebd2a17854c4cfba91b3e800bb66e: Status 404 returned error can't find the container with id f00978733ebd4e9d3763dd20389763b97d4ebd2a17854c4cfba91b3e800bb66e Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.164537 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.164841 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.164897 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz8ld\" (UniqueName: \"kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.164930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.266595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.266631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.266694 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz8ld\" (UniqueName: \"kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.266725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.271805 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.271830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.271844 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.280123 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz8ld\" (UniqueName: \"kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.312848 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.367935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:45 crc kubenswrapper[4824]: E0121 11:25:45.368106 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 21 11:25:45 crc kubenswrapper[4824]: E0121 11:25:45.368121 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 21 11:25:45 crc kubenswrapper[4824]: E0121 11:25:45.368158 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift podName:f980e493-ca74-47f1-8f6f-aca9c9011bf5 nodeName:}" failed. No retries permitted until 2026-01-21 11:25:53.36814564 +0000 UTC m=+895.661174931 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift") pod "swift-storage-0" (UID: "f980e493-ca74-47f1-8f6f-aca9c9011bf5") : configmap "swift-ring-files" not found Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.747316 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sx9lb"] Jan 21 11:25:45 crc kubenswrapper[4824]: W0121 11:25:45.749094 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee7a30d4_dcec_4719_a629_dfaf588d4169.slice/crio-0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c WatchSource:0}: Error finding container 0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c: Status 404 returned error can't find the container with id 0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.767102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nc6vf" event={"ID":"29473359-3517-4b86-bd5e-80e25706ff27","Type":"ContainerStarted","Data":"8b9dec71c98fa449139cdcf21064eb0b0fb8a6897189e13bedcb78be35a24b3d"} Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.771251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sx9lb" event={"ID":"ee7a30d4-dcec-4719-a629-dfaf588d4169","Type":"ContainerStarted","Data":"0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c"} Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.773466 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerID="e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab" exitCode=0 Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.773668 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n665r" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.774251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerDied","Data":"e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab"} Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.774351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerStarted","Data":"f00978733ebd4e9d3763dd20389763b97d4ebd2a17854c4cfba91b3e800bb66e"} Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.787480 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-nc6vf" podStartSLOduration=2.217822141 podStartE2EDuration="4.787461561s" podCreationTimestamp="2026-01-21 11:25:41 +0000 UTC" firstStartedPulling="2026-01-21 11:25:42.196502349 +0000 UTC m=+884.489531642" lastFinishedPulling="2026-01-21 11:25:44.76614177 +0000 UTC m=+887.059171062" observedRunningTime="2026-01-21 11:25:45.783125788 +0000 UTC m=+888.076155080" watchObservedRunningTime="2026-01-21 11:25:45.787461561 +0000 UTC m=+888.080490853" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.087157 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.095275 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.099284 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.179662 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.180011 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.180222 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqdmn\" (UniqueName: \"kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.280989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.281075 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.281174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqdmn\" (UniqueName: \"kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.281742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.281798 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.297090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqdmn\" (UniqueName: \"kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.416675 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.701309 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.766768 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.766937 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="dnsmasq-dns" containerID="cri-o://bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8" gracePeriod=10 Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.793499 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerStarted","Data":"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d"} Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.942415 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:25:46 crc kubenswrapper[4824]: W0121 11:25:46.954562 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6426f379_b3d5_4589_b5e1_659aed932410.slice/crio-bc78d731286fa1edc4820267f7da4e9a25da4496ce988c15be5ce2476d00d97c WatchSource:0}: Error finding container bc78d731286fa1edc4820267f7da4e9a25da4496ce988c15be5ce2476d00d97c: Status 404 returned error can't find the container with id bc78d731286fa1edc4820267f7da4e9a25da4496ce988c15be5ce2476d00d97c Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.246598 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.406412 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcgfn\" (UniqueName: \"kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn\") pod \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.406772 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc\") pod \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.406798 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config\") pod \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.412637 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn" (OuterVolumeSpecName: "kube-api-access-wcgfn") pod "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" (UID: "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4"). InnerVolumeSpecName "kube-api-access-wcgfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.444488 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config" (OuterVolumeSpecName: "config") pod "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" (UID: "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.450888 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" (UID: "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.508404 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcgfn\" (UniqueName: \"kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.508433 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.508442 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.799778 4824 generic.go:334] "Generic (PLEG): container finished" podID="6426f379-b3d5-4589-b5e1-659aed932410" containerID="7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3" exitCode=0 Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.799850 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerDied","Data":"7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3"} Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.799876 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerStarted","Data":"bc78d731286fa1edc4820267f7da4e9a25da4496ce988c15be5ce2476d00d97c"} Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.802020 4824 generic.go:334] "Generic (PLEG): container finished" podID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerID="bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8" exitCode=0 Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.802128 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.802147 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" event={"ID":"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4","Type":"ContainerDied","Data":"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8"} Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.802200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" event={"ID":"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4","Type":"ContainerDied","Data":"9f481192efadf26388108ffc156d5c70e082a7630e4e6aa574c82729d2674bf6"} Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.802219 4824 scope.go:117] "RemoveContainer" containerID="bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.810141 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerID="32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d" exitCode=0 Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.810186 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerDied","Data":"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d"} Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.830775 4824 scope.go:117] "RemoveContainer" containerID="cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.856465 4824 scope.go:117] "RemoveContainer" containerID="bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8" Jan 21 11:25:47 crc kubenswrapper[4824]: E0121 11:25:47.857027 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8\": container with ID starting with bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8 not found: ID does not exist" containerID="bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.857063 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8"} err="failed to get container status \"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8\": rpc error: code = NotFound desc = could not find container \"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8\": container with ID starting with bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8 not found: ID does not exist" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.857084 4824 scope.go:117] "RemoveContainer" containerID="cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059" Jan 21 11:25:47 crc kubenswrapper[4824]: E0121 11:25:47.857430 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059\": container with ID starting with cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059 not found: ID does not exist" containerID="cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.857456 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059"} err="failed to get container status \"cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059\": rpc error: code = NotFound desc = could not find container \"cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059\": container with ID starting with cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059 not found: ID does not exist" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.859414 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.864986 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.000209 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-n665r"] Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.004652 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-n665r"] Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.056567 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" path="/var/lib/kubelet/pods/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4/volumes" Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.057138 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af5c64ab-b4a4-43f5-81c3-416194e40b9a" path="/var/lib/kubelet/pods/af5c64ab-b4a4-43f5-81c3-416194e40b9a/volumes" Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.820156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerStarted","Data":"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279"} Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.822414 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerStarted","Data":"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c"} Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.843196 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gnlcj" podStartSLOduration=3.259282191 podStartE2EDuration="5.843179837s" podCreationTimestamp="2026-01-21 11:25:43 +0000 UTC" firstStartedPulling="2026-01-21 11:25:45.775043593 +0000 UTC m=+888.068072885" lastFinishedPulling="2026-01-21 11:25:48.358941239 +0000 UTC m=+890.651970531" observedRunningTime="2026-01-21 11:25:48.840207994 +0000 UTC m=+891.133237286" watchObservedRunningTime="2026-01-21 11:25:48.843179837 +0000 UTC m=+891.136209129" Jan 21 11:25:49 crc kubenswrapper[4824]: I0121 11:25:49.840837 4824 generic.go:334] "Generic (PLEG): container finished" podID="6426f379-b3d5-4589-b5e1-659aed932410" containerID="44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c" exitCode=0 Jan 21 11:25:49 crc kubenswrapper[4824]: I0121 11:25:49.840993 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerDied","Data":"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c"} Jan 21 11:25:50 crc kubenswrapper[4824]: I0121 11:25:50.848664 4824 generic.go:334] "Generic (PLEG): container finished" podID="29473359-3517-4b86-bd5e-80e25706ff27" containerID="8b9dec71c98fa449139cdcf21064eb0b0fb8a6897189e13bedcb78be35a24b3d" exitCode=0 Jan 21 11:25:50 crc kubenswrapper[4824]: I0121 11:25:50.848772 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nc6vf" event={"ID":"29473359-3517-4b86-bd5e-80e25706ff27","Type":"ContainerDied","Data":"8b9dec71c98fa449139cdcf21064eb0b0fb8a6897189e13bedcb78be35a24b3d"} Jan 21 11:25:50 crc kubenswrapper[4824]: I0121 11:25:50.856625 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerStarted","Data":"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e"} Jan 21 11:25:50 crc kubenswrapper[4824]: I0121 11:25:50.877486 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5s6wj" podStartSLOduration=2.37313704 podStartE2EDuration="4.877472504s" podCreationTimestamp="2026-01-21 11:25:46 +0000 UTC" firstStartedPulling="2026-01-21 11:25:47.801641395 +0000 UTC m=+890.094670688" lastFinishedPulling="2026-01-21 11:25:50.305976861 +0000 UTC m=+892.599006152" observedRunningTime="2026-01-21 11:25:50.87291368 +0000 UTC m=+893.165942973" watchObservedRunningTime="2026-01-21 11:25:50.877472504 +0000 UTC m=+893.170501795" Jan 21 11:25:51 crc kubenswrapper[4824]: I0121 11:25:51.210202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.008700 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-m6rpv"] Jan 21 11:25:53 crc kubenswrapper[4824]: E0121 11:25:53.009206 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="dnsmasq-dns" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.009218 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="dnsmasq-dns" Jan 21 11:25:53 crc kubenswrapper[4824]: E0121 11:25:53.009229 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="init" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.009235 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="init" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.009394 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="dnsmasq-dns" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.009871 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.011619 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.023341 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-m6rpv"] Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.097270 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p2xd\" (UniqueName: \"kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.097437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.200988 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p2xd\" (UniqueName: \"kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.201119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.203508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.223180 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p2xd\" (UniqueName: \"kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.331142 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.403526 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.408515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.494589 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.624121 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.624163 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.658143 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.908016 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:55 crc kubenswrapper[4824]: I0121 11:25:55.476308 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:55 crc kubenswrapper[4824]: I0121 11:25:55.887746 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gnlcj" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="registry-server" containerID="cri-o://39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279" gracePeriod=2 Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.132736 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.254895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.254938 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cffv8\" (UniqueName: \"kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.254993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.255016 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.255046 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.255061 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.255136 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.256536 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.259306 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.261996 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8" (OuterVolumeSpecName: "kube-api-access-cffv8") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "kube-api-access-cffv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.262971 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.274363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.274604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts" (OuterVolumeSpecName: "scripts") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.278934 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.281399 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.356734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mmv9\" (UniqueName: \"kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9\") pod \"bb4180b2-5d3e-486d-8197-348d01d2286e\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.356827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content\") pod \"bb4180b2-5d3e-486d-8197-348d01d2286e\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.356853 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities\") pod \"bb4180b2-5d3e-486d-8197-348d01d2286e\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357309 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357326 4824 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357335 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cffv8\" (UniqueName: \"kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357343 4824 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357351 4824 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357359 4824 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357367 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357930 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities" (OuterVolumeSpecName: "utilities") pod "bb4180b2-5d3e-486d-8197-348d01d2286e" (UID: "bb4180b2-5d3e-486d-8197-348d01d2286e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.359366 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9" (OuterVolumeSpecName: "kube-api-access-6mmv9") pod "bb4180b2-5d3e-486d-8197-348d01d2286e" (UID: "bb4180b2-5d3e-486d-8197-348d01d2286e"). InnerVolumeSpecName "kube-api-access-6mmv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.396377 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-m6rpv"] Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.403339 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb4180b2-5d3e-486d-8197-348d01d2286e" (UID: "bb4180b2-5d3e-486d-8197-348d01d2286e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: W0121 11:25:56.408484 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76e33763_1a41_4640_8e4a_492ab92009e8.slice/crio-78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a WatchSource:0}: Error finding container 78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a: Status 404 returned error can't find the container with id 78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.417830 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.417867 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.459231 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.459406 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.459417 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mmv9\" (UniqueName: \"kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.460433 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.480925 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.493613 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:25:56 crc kubenswrapper[4824]: E0121 11:25:56.493913 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="extract-utilities" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.493934 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="extract-utilities" Jan 21 11:25:56 crc kubenswrapper[4824]: E0121 11:25:56.493949 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29473359-3517-4b86-bd5e-80e25706ff27" containerName="swift-ring-rebalance" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.493968 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="29473359-3517-4b86-bd5e-80e25706ff27" containerName="swift-ring-rebalance" Jan 21 11:25:56 crc kubenswrapper[4824]: E0121 11:25:56.493978 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="extract-content" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.493984 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="extract-content" Jan 21 11:25:56 crc kubenswrapper[4824]: E0121 11:25:56.494006 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="registry-server" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.494012 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="registry-server" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.494145 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="registry-server" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.494163 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="29473359-3517-4b86-bd5e-80e25706ff27" containerName="swift-ring-rebalance" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.496432 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.499309 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.662210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.662340 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd85x\" (UniqueName: \"kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.662497 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.763811 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.763921 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd85x\" (UniqueName: \"kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.764037 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.764311 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.764414 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.781746 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd85x\" (UniqueName: \"kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.829888 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.917313 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.918704 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nc6vf" event={"ID":"29473359-3517-4b86-bd5e-80e25706ff27","Type":"ContainerDied","Data":"be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.918741 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.925326 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sx9lb" event={"ID":"ee7a30d4-dcec-4719-a629-dfaf588d4169","Type":"ContainerStarted","Data":"2b4aa0578cb659bd1fe3cb0351e682b347beae954c60ccf215edaf9c2da1abd4"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.935777 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerID="39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279" exitCode=0 Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.935829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerDied","Data":"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.935852 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerDied","Data":"f00978733ebd4e9d3763dd20389763b97d4ebd2a17854c4cfba91b3e800bb66e"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.935867 4824 scope.go:117] "RemoveContainer" containerID="39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.936242 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.942879 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-sx9lb" podStartSLOduration=2.685972681 podStartE2EDuration="12.942863451s" podCreationTimestamp="2026-01-21 11:25:44 +0000 UTC" firstStartedPulling="2026-01-21 11:25:45.752781961 +0000 UTC m=+888.045811253" lastFinishedPulling="2026-01-21 11:25:56.009672731 +0000 UTC m=+898.302702023" observedRunningTime="2026-01-21 11:25:56.940300941 +0000 UTC m=+899.233330233" watchObservedRunningTime="2026-01-21 11:25:56.942863451 +0000 UTC m=+899.235892743" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.945578 4824 generic.go:334] "Generic (PLEG): container finished" podID="76e33763-1a41-4640-8e4a-492ab92009e8" containerID="25e2c38241236d848b515e556d4747f5cb7cbca9e5eba1926722e4f2e38b424d" exitCode=0 Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.945626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-m6rpv" event={"ID":"76e33763-1a41-4640-8e4a-492ab92009e8","Type":"ContainerDied","Data":"25e2c38241236d848b515e556d4747f5cb7cbca9e5eba1926722e4f2e38b424d"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.945648 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-m6rpv" event={"ID":"76e33763-1a41-4640-8e4a-492ab92009e8","Type":"ContainerStarted","Data":"78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.949051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"5ce1bb7e7ea8646e76b785d5f245fdc4e2efc04e0360410adebd6bb17d781b28"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.990190 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.116500 4824 scope.go:117] "RemoveContainer" containerID="32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.124784 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.131811 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.147709 4824 scope.go:117] "RemoveContainer" containerID="e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.166689 4824 scope.go:117] "RemoveContainer" containerID="39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279" Jan 21 11:25:57 crc kubenswrapper[4824]: E0121 11:25:57.167010 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279\": container with ID starting with 39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279 not found: ID does not exist" containerID="39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.167048 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279"} err="failed to get container status \"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279\": rpc error: code = NotFound desc = could not find container \"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279\": container with ID starting with 39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279 not found: ID does not exist" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.167073 4824 scope.go:117] "RemoveContainer" containerID="32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d" Jan 21 11:25:57 crc kubenswrapper[4824]: E0121 11:25:57.167370 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d\": container with ID starting with 32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d not found: ID does not exist" containerID="32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.167392 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d"} err="failed to get container status \"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d\": rpc error: code = NotFound desc = could not find container \"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d\": container with ID starting with 32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d not found: ID does not exist" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.167415 4824 scope.go:117] "RemoveContainer" containerID="e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab" Jan 21 11:25:57 crc kubenswrapper[4824]: E0121 11:25:57.169857 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab\": container with ID starting with e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab not found: ID does not exist" containerID="e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.169887 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab"} err="failed to get container status \"e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab\": rpc error: code = NotFound desc = could not find container \"e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab\": container with ID starting with e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab not found: ID does not exist" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.231495 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:25:57 crc kubenswrapper[4824]: W0121 11:25:57.238445 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c75f0f9_53e1_4a64_9211_559ef51c1b53.slice/crio-799120a9a7a20b359b616c2030579c280c96e8a87a304704ae580c5acba60e25 WatchSource:0}: Error finding container 799120a9a7a20b359b616c2030579c280c96e8a87a304704ae580c5acba60e25: Status 404 returned error can't find the container with id 799120a9a7a20b359b616c2030579c280c96e8a87a304704ae580c5acba60e25 Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.956876 4824 generic.go:334] "Generic (PLEG): container finished" podID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerID="12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17" exitCode=0 Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.956906 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerDied","Data":"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17"} Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.958354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerStarted","Data":"799120a9a7a20b359b616c2030579c280c96e8a87a304704ae580c5acba60e25"} Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.959993 4824 generic.go:334] "Generic (PLEG): container finished" podID="156fec50-b486-4e84-a7bf-b40491a863c7" containerID="57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487" exitCode=0 Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.960020 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerDied","Data":"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487"} Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.062585 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" path="/var/lib/kubelet/pods/bb4180b2-5d3e-486d-8197-348d01d2286e/volumes" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.213108 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.292555 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts\") pod \"76e33763-1a41-4640-8e4a-492ab92009e8\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.292655 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p2xd\" (UniqueName: \"kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd\") pod \"76e33763-1a41-4640-8e4a-492ab92009e8\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.293391 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "76e33763-1a41-4640-8e4a-492ab92009e8" (UID: "76e33763-1a41-4640-8e4a-492ab92009e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.296130 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd" (OuterVolumeSpecName: "kube-api-access-2p2xd") pod "76e33763-1a41-4640-8e4a-492ab92009e8" (UID: "76e33763-1a41-4640-8e4a-492ab92009e8"). InnerVolumeSpecName "kube-api-access-2p2xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.394122 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.394148 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p2xd\" (UniqueName: \"kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.874700 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.967370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerStarted","Data":"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32"} Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.968081 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.969629 4824 generic.go:334] "Generic (PLEG): container finished" podID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerID="dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76" exitCode=0 Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.969684 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerDied","Data":"dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76"} Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.971276 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-m6rpv" event={"ID":"76e33763-1a41-4640-8e4a-492ab92009e8","Type":"ContainerDied","Data":"78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a"} Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.971293 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.971330 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.984489 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5s6wj" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="registry-server" containerID="cri-o://828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e" gracePeriod=2 Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.984737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerStarted","Data":"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7"} Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.985519 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.986804 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=43.4561252 podStartE2EDuration="48.98679663s" podCreationTimestamp="2026-01-21 11:25:10 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.035726855 +0000 UTC m=+861.328756147" lastFinishedPulling="2026-01-21 11:25:24.566398285 +0000 UTC m=+866.859427577" observedRunningTime="2026-01-21 11:25:58.985856247 +0000 UTC m=+901.278885540" watchObservedRunningTime="2026-01-21 11:25:58.98679663 +0000 UTC m=+901.279825922" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.011445 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=43.567915502 podStartE2EDuration="49.011429051s" podCreationTimestamp="2026-01-21 11:25:10 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.153926917 +0000 UTC m=+861.446956210" lastFinishedPulling="2026-01-21 11:25:24.597440467 +0000 UTC m=+866.890469759" observedRunningTime="2026-01-21 11:25:59.006029493 +0000 UTC m=+901.299058785" watchObservedRunningTime="2026-01-21 11:25:59.011429051 +0000 UTC m=+901.304458343" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.382541 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.509785 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content\") pod \"6426f379-b3d5-4589-b5e1-659aed932410\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.509990 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities\") pod \"6426f379-b3d5-4589-b5e1-659aed932410\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.510036 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqdmn\" (UniqueName: \"kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn\") pod \"6426f379-b3d5-4589-b5e1-659aed932410\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.510574 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities" (OuterVolumeSpecName: "utilities") pod "6426f379-b3d5-4589-b5e1-659aed932410" (UID: "6426f379-b3d5-4589-b5e1-659aed932410"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.515280 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn" (OuterVolumeSpecName: "kube-api-access-rqdmn") pod "6426f379-b3d5-4589-b5e1-659aed932410" (UID: "6426f379-b3d5-4589-b5e1-659aed932410"). InnerVolumeSpecName "kube-api-access-rqdmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.612103 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.612125 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqdmn\" (UniqueName: \"kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.695813 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6426f379-b3d5-4589-b5e1-659aed932410" (UID: "6426f379-b3d5-4589-b5e1-659aed932410"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.713331 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.991645 4824 generic.go:334] "Generic (PLEG): container finished" podID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerID="af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71" exitCode=0 Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.991723 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerDied","Data":"af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.993793 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"d29ad988a899f06204abf43bd5ea374c7b88de2de25e37348d02940d02652131"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.993996 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"b804a5ae6f06ff30e4552c437c35633b5fa1b97571607261f6b664323dc9ffc2"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.994009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"857de0e92541a2e1927ec6b1d7b50def1b5a78ae48f2a6eb7c779e2fc4c4b45e"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.994016 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"aaccd44ac3f4aff508890cdda80001df1cc8ec7e9277bfe539d8bef3c06d1be3"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.995805 4824 generic.go:334] "Generic (PLEG): container finished" podID="6426f379-b3d5-4589-b5e1-659aed932410" containerID="828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e" exitCode=0 Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.996021 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerDied","Data":"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.996045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerDied","Data":"bc78d731286fa1edc4820267f7da4e9a25da4496ce988c15be5ce2476d00d97c"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.996064 4824 scope.go:117] "RemoveContainer" containerID="828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.996079 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.011514 4824 scope.go:117] "RemoveContainer" containerID="44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.029306 4824 scope.go:117] "RemoveContainer" containerID="7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.030940 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.036948 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.053488 4824 scope.go:117] "RemoveContainer" containerID="828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.055689 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e\": container with ID starting with 828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e not found: ID does not exist" containerID="828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.055719 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e"} err="failed to get container status \"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e\": rpc error: code = NotFound desc = could not find container \"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e\": container with ID starting with 828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e not found: ID does not exist" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.055736 4824 scope.go:117] "RemoveContainer" containerID="44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.056093 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c\": container with ID starting with 44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c not found: ID does not exist" containerID="44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.056127 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c"} err="failed to get container status \"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c\": rpc error: code = NotFound desc = could not find container \"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c\": container with ID starting with 44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c not found: ID does not exist" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.056149 4824 scope.go:117] "RemoveContainer" containerID="7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.056450 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3\": container with ID starting with 7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3 not found: ID does not exist" containerID="7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.056475 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3"} err="failed to get container status \"7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3\": rpc error: code = NotFound desc = could not find container \"7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3\": container with ID starting with 7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3 not found: ID does not exist" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.062658 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6426f379-b3d5-4589-b5e1-659aed932410" path="/var/lib/kubelet/pods/6426f379-b3d5-4589-b5e1-659aed932410/volumes" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.137654 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-htwd6" podUID="511f2b6d-a08d-49f8-b393-ab222219d4a7" containerName="ovn-controller" probeResult="failure" output=< Jan 21 11:26:00 crc kubenswrapper[4824]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 21 11:26:00 crc kubenswrapper[4824]: > Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.242468 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.257166 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448274 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-htwd6-config-jbc69"] Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.448576 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="registry-server" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448592 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="registry-server" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.448604 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="extract-content" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448610 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="extract-content" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.448619 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e33763-1a41-4640-8e4a-492ab92009e8" containerName="mariadb-account-create-update" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448624 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e33763-1a41-4640-8e4a-492ab92009e8" containerName="mariadb-account-create-update" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.448642 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="extract-utilities" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448648 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="extract-utilities" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448806 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e33763-1a41-4640-8e4a-492ab92009e8" containerName="mariadb-account-create-update" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448826 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="registry-server" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.449314 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.456026 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.469028 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-htwd6-config-jbc69"] Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.533784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.533818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmxtd\" (UniqueName: \"kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.533870 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.533998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.534022 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.534053 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635451 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmxtd\" (UniqueName: \"kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.636068 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.636087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.636253 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.636430 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.637474 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.658183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmxtd\" (UniqueName: \"kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.773205 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:01 crc kubenswrapper[4824]: I0121 11:26:01.006021 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"dc6ddaa09fb1ec22ec741c78662d75b99df4b6da66737456c8dccf858b7522d7"} Jan 21 11:26:01 crc kubenswrapper[4824]: I0121 11:26:01.006398 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"4bda5ab0e896dbb529ef5290ea2a3c7e17af838da67d703a146d1016845d42a6"} Jan 21 11:26:01 crc kubenswrapper[4824]: I0121 11:26:01.011202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerStarted","Data":"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45"} Jan 21 11:26:01 crc kubenswrapper[4824]: I0121 11:26:01.028754 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b795q" podStartSLOduration=3.321225141 podStartE2EDuration="5.028740899s" podCreationTimestamp="2026-01-21 11:25:56 +0000 UTC" firstStartedPulling="2026-01-21 11:25:58.970461649 +0000 UTC m=+901.263490941" lastFinishedPulling="2026-01-21 11:26:00.677977407 +0000 UTC m=+902.971006699" observedRunningTime="2026-01-21 11:26:01.024055567 +0000 UTC m=+903.317084870" watchObservedRunningTime="2026-01-21 11:26:01.028740899 +0000 UTC m=+903.321770181" Jan 21 11:26:01 crc kubenswrapper[4824]: I0121 11:26:01.149091 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-htwd6-config-jbc69"] Jan 21 11:26:01 crc kubenswrapper[4824]: W0121 11:26:01.159577 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d3982e6_513d_454f_a08a_34b32779b559.slice/crio-f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712 WatchSource:0}: Error finding container f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712: Status 404 returned error can't find the container with id f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712 Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.021934 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"e964792da3bebb08ca73e92297f79f1b50c4c3b30e75a5c6fc4c9f54326727d1"} Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.022264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"a2eb36fe6d265a82cf21727e4989a133a77f37bcc8b16fed1124388f79b0884b"} Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.023784 4824 generic.go:334] "Generic (PLEG): container finished" podID="6d3982e6-513d-454f-a08a-34b32779b559" containerID="7802d3564cb8577bdb2e0cb7acab4d7e0721818cb3c2ece34e851778aff1034e" exitCode=0 Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.023827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-htwd6-config-jbc69" event={"ID":"6d3982e6-513d-454f-a08a-34b32779b559","Type":"ContainerDied","Data":"7802d3564cb8577bdb2e0cb7acab4d7e0721818cb3c2ece34e851778aff1034e"} Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.023843 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-htwd6-config-jbc69" event={"ID":"6d3982e6-513d-454f-a08a-34b32779b559","Type":"ContainerStarted","Data":"f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712"} Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.025421 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee7a30d4-dcec-4719-a629-dfaf588d4169" containerID="2b4aa0578cb659bd1fe3cb0351e682b347beae954c60ccf215edaf9c2da1abd4" exitCode=0 Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.026137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sx9lb" event={"ID":"ee7a30d4-dcec-4719-a629-dfaf588d4169","Type":"ContainerDied","Data":"2b4aa0578cb659bd1fe3cb0351e682b347beae954c60ccf215edaf9c2da1abd4"} Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.034193 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"3455e0eeedb7b32a4c23c322afa378a52b1af730fe411dd14be91d06eaa924c6"} Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.034398 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"c626637d2ac5daec740897c7b85b3e7ac1dcc9b2bf37eef13e76aae43abc2677"} Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.034410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"b954af21a9cbab6f49f0ca38e4aa8205d0f06bb458256451df49256caf0cec2e"} Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.286286 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373101 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373145 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373186 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmxtd\" (UniqueName: \"kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373280 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373343 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373364 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373455 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run" (OuterVolumeSpecName: "var-run") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373767 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373935 4824 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373953 4824 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373973 4824 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373983 4824 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.374079 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts" (OuterVolumeSpecName: "scripts") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.377917 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd" (OuterVolumeSpecName: "kube-api-access-dmxtd") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "kube-api-access-dmxtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.386830 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sx9lb" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.474666 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle\") pod \"ee7a30d4-dcec-4719-a629-dfaf588d4169\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.474723 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data\") pod \"ee7a30d4-dcec-4719-a629-dfaf588d4169\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.474807 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz8ld\" (UniqueName: \"kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld\") pod \"ee7a30d4-dcec-4719-a629-dfaf588d4169\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.474874 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data\") pod \"ee7a30d4-dcec-4719-a629-dfaf588d4169\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.475234 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmxtd\" (UniqueName: \"kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.475253 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.477983 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ee7a30d4-dcec-4719-a629-dfaf588d4169" (UID: "ee7a30d4-dcec-4719-a629-dfaf588d4169"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.478249 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld" (OuterVolumeSpecName: "kube-api-access-cz8ld") pod "ee7a30d4-dcec-4719-a629-dfaf588d4169" (UID: "ee7a30d4-dcec-4719-a629-dfaf588d4169"). InnerVolumeSpecName "kube-api-access-cz8ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.490604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee7a30d4-dcec-4719-a629-dfaf588d4169" (UID: "ee7a30d4-dcec-4719-a629-dfaf588d4169"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.502760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data" (OuterVolumeSpecName: "config-data") pod "ee7a30d4-dcec-4719-a629-dfaf588d4169" (UID: "ee7a30d4-dcec-4719-a629-dfaf588d4169"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.576292 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz8ld\" (UniqueName: \"kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.576318 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.576329 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.576337 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.048248 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sx9lb" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.048242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sx9lb" event={"ID":"ee7a30d4-dcec-4719-a629-dfaf588d4169","Type":"ContainerDied","Data":"0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c"} Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.048713 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.051074 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.060524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-htwd6-config-jbc69" event={"ID":"6d3982e6-513d-454f-a08a-34b32779b559","Type":"ContainerDied","Data":"f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712"} Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.060558 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.384680 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:04 crc kubenswrapper[4824]: E0121 11:26:04.384967 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7a30d4-dcec-4719-a629-dfaf588d4169" containerName="glance-db-sync" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.384995 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7a30d4-dcec-4719-a629-dfaf588d4169" containerName="glance-db-sync" Jan 21 11:26:04 crc kubenswrapper[4824]: E0121 11:26:04.385015 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d3982e6-513d-454f-a08a-34b32779b559" containerName="ovn-config" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.385020 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d3982e6-513d-454f-a08a-34b32779b559" containerName="ovn-config" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.385164 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7a30d4-dcec-4719-a629-dfaf588d4169" containerName="glance-db-sync" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.385177 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d3982e6-513d-454f-a08a-34b32779b559" containerName="ovn-config" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.385851 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.402086 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.454610 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-htwd6-config-jbc69"] Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.463637 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-htwd6-config-jbc69"] Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.487065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.487146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.487224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpg89\" (UniqueName: \"kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.487265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.487320 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.588995 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpg89\" (UniqueName: \"kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.589058 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.589113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.589156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.589215 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.589987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.590009 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.590094 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.590102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.605356 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpg89\" (UniqueName: \"kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.701243 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.061500 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"19a1cf9e015e25e337fcd599f29b0b6c4f318463d25fc0a8e8b038d88277f73f"} Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.061711 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"8f762056829510d59deced7738a3ca721ea70f910b73bcb4d792931d7b694216"} Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.061722 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"ebfa28688dafa5bd49a43d8c94f03fab1ec7dfa1926981e4bcf2dade062d2ccf"} Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.061729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"6c4e2b6ccc2b96fd74c1d55a4f0ff2eb7064ee3cd7ca9e89713240c4c149439d"} Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.079514 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.087361 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=23.215689231 podStartE2EDuration="29.087346411s" podCreationTimestamp="2026-01-21 11:25:36 +0000 UTC" firstStartedPulling="2026-01-21 11:25:56.493180239 +0000 UTC m=+898.786209531" lastFinishedPulling="2026-01-21 11:26:02.364837419 +0000 UTC m=+904.657866711" observedRunningTime="2026-01-21 11:26:05.083921436 +0000 UTC m=+907.376950727" watchObservedRunningTime="2026-01-21 11:26:05.087346411 +0000 UTC m=+907.380375703" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.136778 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-htwd6" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.366966 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.400993 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.402027 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.404851 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.419089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504341 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504537 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504597 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504639 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504781 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xgrw\" (UniqueName: \"kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504820 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606283 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xgrw\" (UniqueName: \"kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.607191 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.607190 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.607234 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.607338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.607902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.620170 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xgrw\" (UniqueName: \"kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.717566 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.056633 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d3982e6-513d-454f-a08a-34b32779b559" path="/var/lib/kubelet/pods/6d3982e6-513d-454f-a08a-34b32779b559/volumes" Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.068267 4824 generic.go:334] "Generic (PLEG): container finished" podID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerID="a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f" exitCode=0 Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.068362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" event={"ID":"3e491cbf-b795-4735-a356-c30d79d2fefa","Type":"ContainerDied","Data":"a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f"} Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.068401 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" event={"ID":"3e491cbf-b795-4735-a356-c30d79d2fefa","Type":"ContainerStarted","Data":"c2adb67835a6dab5ee88f8c74c3273f9d6eb5ecef5ab9da3b16f2dada3d6a2a1"} Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.115876 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.830501 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.830701 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.863239 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.074564 4824 generic.go:334] "Generic (PLEG): container finished" podID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerID="0aacefa5ca6057f616f2afc320e43a0e95b25311f579f1cc69ade834fb63b2f0" exitCode=0 Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.074629 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" event={"ID":"c4a81903-5e1f-4d9c-bacf-60639069db18","Type":"ContainerDied","Data":"0aacefa5ca6057f616f2afc320e43a0e95b25311f579f1cc69ade834fb63b2f0"} Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.074654 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" event={"ID":"c4a81903-5e1f-4d9c-bacf-60639069db18","Type":"ContainerStarted","Data":"19d64cd123fe9ef3646e70cf5808fefa9acba69592569cd21640e9399cbca5d9"} Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.076133 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" event={"ID":"3e491cbf-b795-4735-a356-c30d79d2fefa","Type":"ContainerStarted","Data":"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f"} Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.076210 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="dnsmasq-dns" containerID="cri-o://481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f" gracePeriod=10 Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.076707 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.102921 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" podStartSLOduration=3.102904885 podStartE2EDuration="3.102904885s" podCreationTimestamp="2026-01-21 11:26:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:07.101426197 +0000 UTC m=+909.394455489" watchObservedRunningTime="2026-01-21 11:26:07.102904885 +0000 UTC m=+909.395934176" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.110598 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.158504 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.428407 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.534510 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config\") pod \"3e491cbf-b795-4735-a356-c30d79d2fefa\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.534616 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb\") pod \"3e491cbf-b795-4735-a356-c30d79d2fefa\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.534646 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc\") pod \"3e491cbf-b795-4735-a356-c30d79d2fefa\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.534688 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb\") pod \"3e491cbf-b795-4735-a356-c30d79d2fefa\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.534780 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpg89\" (UniqueName: \"kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89\") pod \"3e491cbf-b795-4735-a356-c30d79d2fefa\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.538042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89" (OuterVolumeSpecName: "kube-api-access-fpg89") pod "3e491cbf-b795-4735-a356-c30d79d2fefa" (UID: "3e491cbf-b795-4735-a356-c30d79d2fefa"). InnerVolumeSpecName "kube-api-access-fpg89". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.563485 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config" (OuterVolumeSpecName: "config") pod "3e491cbf-b795-4735-a356-c30d79d2fefa" (UID: "3e491cbf-b795-4735-a356-c30d79d2fefa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.563804 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3e491cbf-b795-4735-a356-c30d79d2fefa" (UID: "3e491cbf-b795-4735-a356-c30d79d2fefa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.568488 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e491cbf-b795-4735-a356-c30d79d2fefa" (UID: "3e491cbf-b795-4735-a356-c30d79d2fefa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.569377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e491cbf-b795-4735-a356-c30d79d2fefa" (UID: "3e491cbf-b795-4735-a356-c30d79d2fefa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.636795 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.637036 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.637116 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.637178 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.637240 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpg89\" (UniqueName: \"kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.083725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" event={"ID":"c4a81903-5e1f-4d9c-bacf-60639069db18","Type":"ContainerStarted","Data":"862050906c3faeb783c3cf1a55874b97964bf3e73dcab5972c6e49f1ea08c530"} Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.083787 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.086124 4824 generic.go:334] "Generic (PLEG): container finished" podID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerID="481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f" exitCode=0 Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.086179 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.086213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" event={"ID":"3e491cbf-b795-4735-a356-c30d79d2fefa","Type":"ContainerDied","Data":"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f"} Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.086237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" event={"ID":"3e491cbf-b795-4735-a356-c30d79d2fefa","Type":"ContainerDied","Data":"c2adb67835a6dab5ee88f8c74c3273f9d6eb5ecef5ab9da3b16f2dada3d6a2a1"} Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.086253 4824 scope.go:117] "RemoveContainer" containerID="481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.101217 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" podStartSLOduration=3.101204052 podStartE2EDuration="3.101204052s" podCreationTimestamp="2026-01-21 11:26:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:08.099172212 +0000 UTC m=+910.392201505" watchObservedRunningTime="2026-01-21 11:26:08.101204052 +0000 UTC m=+910.394233344" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.106224 4824 scope.go:117] "RemoveContainer" containerID="a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.122526 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.126647 4824 scope.go:117] "RemoveContainer" containerID="481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f" Jan 21 11:26:08 crc kubenswrapper[4824]: E0121 11:26:08.127062 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f\": container with ID starting with 481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f not found: ID does not exist" containerID="481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.127092 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f"} err="failed to get container status \"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f\": rpc error: code = NotFound desc = could not find container \"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f\": container with ID starting with 481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f not found: ID does not exist" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.127112 4824 scope.go:117] "RemoveContainer" containerID="a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f" Jan 21 11:26:08 crc kubenswrapper[4824]: E0121 11:26:08.127343 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f\": container with ID starting with a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f not found: ID does not exist" containerID="a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.127361 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f"} err="failed to get container status \"a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f\": rpc error: code = NotFound desc = could not find container \"a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f\": container with ID starting with a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f not found: ID does not exist" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.128100 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.093170 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b795q" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="registry-server" containerID="cri-o://9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45" gracePeriod=2 Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.470513 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.566771 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities\") pod \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.566834 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content\") pod \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.566913 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd85x\" (UniqueName: \"kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x\") pod \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.567443 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities" (OuterVolumeSpecName: "utilities") pod "1c75f0f9-53e1-4a64-9211-559ef51c1b53" (UID: "1c75f0f9-53e1-4a64-9211-559ef51c1b53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.571399 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x" (OuterVolumeSpecName: "kube-api-access-hd85x") pod "1c75f0f9-53e1-4a64-9211-559ef51c1b53" (UID: "1c75f0f9-53e1-4a64-9211-559ef51c1b53"). InnerVolumeSpecName "kube-api-access-hd85x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.582568 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c75f0f9-53e1-4a64-9211-559ef51c1b53" (UID: "1c75f0f9-53e1-4a64-9211-559ef51c1b53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.668741 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.668775 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.668786 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd85x\" (UniqueName: \"kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.056572 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" path="/var/lib/kubelet/pods/3e491cbf-b795-4735-a356-c30d79d2fefa/volumes" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.100338 4824 generic.go:334] "Generic (PLEG): container finished" podID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerID="9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45" exitCode=0 Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.100373 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerDied","Data":"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45"} Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.100383 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.100396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerDied","Data":"799120a9a7a20b359b616c2030579c280c96e8a87a304704ae580c5acba60e25"} Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.100411 4824 scope.go:117] "RemoveContainer" containerID="9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.115825 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.117094 4824 scope.go:117] "RemoveContainer" containerID="af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.120579 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.133688 4824 scope.go:117] "RemoveContainer" containerID="dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.155401 4824 scope.go:117] "RemoveContainer" containerID="9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45" Jan 21 11:26:10 crc kubenswrapper[4824]: E0121 11:26:10.155821 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45\": container with ID starting with 9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45 not found: ID does not exist" containerID="9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.155857 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45"} err="failed to get container status \"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45\": rpc error: code = NotFound desc = could not find container \"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45\": container with ID starting with 9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45 not found: ID does not exist" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.155880 4824 scope.go:117] "RemoveContainer" containerID="af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71" Jan 21 11:26:10 crc kubenswrapper[4824]: E0121 11:26:10.156222 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71\": container with ID starting with af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71 not found: ID does not exist" containerID="af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.156262 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71"} err="failed to get container status \"af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71\": rpc error: code = NotFound desc = could not find container \"af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71\": container with ID starting with af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71 not found: ID does not exist" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.156289 4824 scope.go:117] "RemoveContainer" containerID="dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76" Jan 21 11:26:10 crc kubenswrapper[4824]: E0121 11:26:10.156566 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76\": container with ID starting with dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76 not found: ID does not exist" containerID="dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.156593 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76"} err="failed to get container status \"dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76\": rpc error: code = NotFound desc = could not find container \"dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76\": container with ID starting with dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76 not found: ID does not exist" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.376139 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.633114 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812292 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-kk8kb"] Jan 21 11:26:11 crc kubenswrapper[4824]: E0121 11:26:11.812557 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="extract-utilities" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812574 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="extract-utilities" Jan 21 11:26:11 crc kubenswrapper[4824]: E0121 11:26:11.812584 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="registry-server" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812591 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="registry-server" Jan 21 11:26:11 crc kubenswrapper[4824]: E0121 11:26:11.812610 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="extract-content" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812615 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="extract-content" Jan 21 11:26:11 crc kubenswrapper[4824]: E0121 11:26:11.812626 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="init" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812631 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="init" Jan 21 11:26:11 crc kubenswrapper[4824]: E0121 11:26:11.812640 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="dnsmasq-dns" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812645 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="dnsmasq-dns" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812809 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="registry-server" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812821 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="dnsmasq-dns" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.813267 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.823549 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-2619-account-create-update-9csh2"] Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.824376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.827914 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.835662 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-kk8kb"] Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.840730 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-2619-account-create-update-9csh2"] Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.902420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.902571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.902663 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4p9x\" (UniqueName: \"kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.902742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.944943 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-84f4b"] Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.945779 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.949190 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.950121 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.950258 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxhzv" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.950274 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.986805 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-84f4b"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.003885 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.003930 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4p9x\" (UniqueName: \"kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.003973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004094 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4nqn\" (UniqueName: \"kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004212 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004230 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004706 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004796 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.008997 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-twgjw"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.009915 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.018858 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-twgjw"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.018902 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-3c9f-account-create-update-rtprr"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.019887 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.024300 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4p9x\" (UniqueName: \"kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.025826 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3c9f-account-create-update-rtprr"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.033340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.035068 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.056496 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" path="/var/lib/kubelet/pods/1c75f0f9-53e1-4a64-9211-559ef51c1b53/volumes" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.092231 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-ngtnj"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.093373 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.097834 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-ngtnj"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.111860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112027 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbmg7\" (UniqueName: \"kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112127 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6876j\" (UniqueName: \"kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112162 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112217 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4nqn\" (UniqueName: \"kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.117272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.117554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.126269 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4nqn\" (UniqueName: \"kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.134486 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.140859 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.199732 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-60f6-account-create-update-pjh6n"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.200767 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.202913 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.207060 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-nskxc"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.211198 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.215725 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-60f6-account-create-update-pjh6n"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216617 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6876j\" (UniqueName: \"kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216686 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nfhq\" (UniqueName: \"kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216803 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbmg7\" (UniqueName: \"kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.217832 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.219408 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.245022 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6876j\" (UniqueName: \"kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.246442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbmg7\" (UniqueName: \"kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.260502 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-nskxc"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.266240 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.314222 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-0662-account-create-update-j7bhp"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.315444 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.317209 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.317732 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.317831 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq7qd\" (UniqueName: \"kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.317861 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqhbb\" (UniqueName: \"kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.317986 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.318021 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.318050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nfhq\" (UniqueName: \"kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.325057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.329633 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-0662-account-create-update-j7bhp"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.346690 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nfhq\" (UniqueName: \"kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.380341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.385489 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.412833 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvj2q\" (UniqueName: \"kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq7qd\" (UniqueName: \"kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqhbb\" (UniqueName: \"kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420773 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.421488 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.421496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.435515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqhbb\" (UniqueName: \"kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.436950 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq7qd\" (UniqueName: \"kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.516744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.521865 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvj2q\" (UniqueName: \"kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.521980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.522588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.534472 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvj2q\" (UniqueName: \"kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.539459 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.601502 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-kk8kb"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.637035 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.687867 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-2619-account-create-update-9csh2"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.982480 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-84f4b"] Jan 21 11:26:12 crc kubenswrapper[4824]: W0121 11:26:12.986526 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5240d35b_cbf1_472d_91da_debf418dd208.slice/crio-9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581 WatchSource:0}: Error finding container 9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581: Status 404 returned error can't find the container with id 9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581 Jan 21 11:26:13 crc kubenswrapper[4824]: E0121 11:26:13.005760 4824 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.26.120:34444->192.168.26.120:32865: write tcp 192.168.26.120:34444->192.168.26.120:32865: write: broken pipe Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.135203 4824 generic.go:334] "Generic (PLEG): container finished" podID="8600373d-f466-43d4-92fc-1fa938f6e91b" containerID="81f52e55093614da3aed12f82750111917b47ee8401eb58ccc0da58488e3019f" exitCode=0 Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.135419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kk8kb" event={"ID":"8600373d-f466-43d4-92fc-1fa938f6e91b","Type":"ContainerDied","Data":"81f52e55093614da3aed12f82750111917b47ee8401eb58ccc0da58488e3019f"} Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.135443 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kk8kb" event={"ID":"8600373d-f466-43d4-92fc-1fa938f6e91b","Type":"ContainerStarted","Data":"e76ccb54eba21f586b5ec6a78b2a5550037daa9208d52611b08df26c0e3ac56b"} Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.135601 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-nskxc"] Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.142010 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3c9f-account-create-update-rtprr"] Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.143146 4824 generic.go:334] "Generic (PLEG): container finished" podID="fe6e7565-793b-4555-b6d5-758f87a5b9c7" containerID="f658b544252ce5134cf2b3d03b008e5a8cf0c0a33b02b6a5554a5288bb79bbfe" exitCode=0 Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.143203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2619-account-create-update-9csh2" event={"ID":"fe6e7565-793b-4555-b6d5-758f87a5b9c7","Type":"ContainerDied","Data":"f658b544252ce5134cf2b3d03b008e5a8cf0c0a33b02b6a5554a5288bb79bbfe"} Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.143225 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2619-account-create-update-9csh2" event={"ID":"fe6e7565-793b-4555-b6d5-758f87a5b9c7","Type":"ContainerStarted","Data":"8526f4c3098ff09c46e3c8957568e817f203fa1293f2cd973998f5c043bb31d2"} Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.145689 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-84f4b" event={"ID":"5240d35b-cbf1-472d-91da-debf418dd208","Type":"ContainerStarted","Data":"9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581"} Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.146934 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-60f6-account-create-update-pjh6n"] Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.222149 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-twgjw"] Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.333620 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-ngtnj"] Jan 21 11:26:13 crc kubenswrapper[4824]: W0121 11:26:13.339827 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9be37b35_7ad5_434a_8d16_c2f9f1661821.slice/crio-e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822 WatchSource:0}: Error finding container e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822: Status 404 returned error can't find the container with id e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822 Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.350501 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-0662-account-create-update-j7bhp"] Jan 21 11:26:13 crc kubenswrapper[4824]: W0121 11:26:13.384667 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4773e7a_a50c_42e6_bb27_a25b0055b8e3.slice/crio-6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf WatchSource:0}: Error finding container 6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf: Status 404 returned error can't find the container with id 6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.167476 4824 generic.go:334] "Generic (PLEG): container finished" podID="b4773e7a-a50c-42e6-bb27-a25b0055b8e3" containerID="cc788777a506818eef849722d9795fcf44c54eb36cb8cfbc69470609fce80bd0" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.167725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0662-account-create-update-j7bhp" event={"ID":"b4773e7a-a50c-42e6-bb27-a25b0055b8e3","Type":"ContainerDied","Data":"cc788777a506818eef849722d9795fcf44c54eb36cb8cfbc69470609fce80bd0"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.167748 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0662-account-create-update-j7bhp" event={"ID":"b4773e7a-a50c-42e6-bb27-a25b0055b8e3","Type":"ContainerStarted","Data":"6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.171745 4824 generic.go:334] "Generic (PLEG): container finished" podID="a10a8f17-c27b-4544-8f9d-a6afd9991a8e" containerID="b6f52bcc83bf92c2ef07afe3fa11fe445d957e311fc35ba4f1767a43394658f9" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.171807 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nskxc" event={"ID":"a10a8f17-c27b-4544-8f9d-a6afd9991a8e","Type":"ContainerDied","Data":"b6f52bcc83bf92c2ef07afe3fa11fe445d957e311fc35ba4f1767a43394658f9"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.171831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nskxc" event={"ID":"a10a8f17-c27b-4544-8f9d-a6afd9991a8e","Type":"ContainerStarted","Data":"d2e64d054f528de025088752cbba2a64296b9807a6337f5fdd775173aefc9828"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.178982 4824 generic.go:334] "Generic (PLEG): container finished" podID="93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" containerID="45d2f067a75588f2f74c579f26d8f9af2b6a872de4eb9d340b34e9fee4726547" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.179101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-60f6-account-create-update-pjh6n" event={"ID":"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c","Type":"ContainerDied","Data":"45d2f067a75588f2f74c579f26d8f9af2b6a872de4eb9d340b34e9fee4726547"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.179163 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-60f6-account-create-update-pjh6n" event={"ID":"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c","Type":"ContainerStarted","Data":"a60e0d515d078f64392d5d1ca5e1034d522ce6ad7bf98f9a82fa5e036ed775d2"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.184371 4824 generic.go:334] "Generic (PLEG): container finished" podID="5c0ddc66-7696-4816-8e96-340aeb39d242" containerID="2176bc084f2009e4d942790b8c3474f286b639f648b347b28da26c0a5a3ccec2" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.184430 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3c9f-account-create-update-rtprr" event={"ID":"5c0ddc66-7696-4816-8e96-340aeb39d242","Type":"ContainerDied","Data":"2176bc084f2009e4d942790b8c3474f286b639f648b347b28da26c0a5a3ccec2"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.184453 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3c9f-account-create-update-rtprr" event={"ID":"5c0ddc66-7696-4816-8e96-340aeb39d242","Type":"ContainerStarted","Data":"e9e93c5745e7fe74ef1e643164e9a0a4e165267deeb109faf4df81c36f336691"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.194506 4824 generic.go:334] "Generic (PLEG): container finished" podID="9be37b35-7ad5-434a-8d16-c2f9f1661821" containerID="c1a92fb4372889a2ab5383adea30e43a1fa74caae3039082b2ef5482212d2758" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.194557 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ngtnj" event={"ID":"9be37b35-7ad5-434a-8d16-c2f9f1661821","Type":"ContainerDied","Data":"c1a92fb4372889a2ab5383adea30e43a1fa74caae3039082b2ef5482212d2758"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.194575 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ngtnj" event={"ID":"9be37b35-7ad5-434a-8d16-c2f9f1661821","Type":"ContainerStarted","Data":"e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.209338 4824 generic.go:334] "Generic (PLEG): container finished" podID="7b747df8-e61f-4689-a08c-0dc6688f1a0e" containerID="e12a5e35428ef122a88a06ec7747afe8e712c9a7807b731859f6ec595579f719" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.209393 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-twgjw" event={"ID":"7b747df8-e61f-4689-a08c-0dc6688f1a0e","Type":"ContainerDied","Data":"e12a5e35428ef122a88a06ec7747afe8e712c9a7807b731859f6ec595579f719"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.209425 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-twgjw" event={"ID":"7b747df8-e61f-4689-a08c-0dc6688f1a0e","Type":"ContainerStarted","Data":"5cbfd6e560404567a0bbc9ae88f4245a4c606e0f65ba74f1fdb8d8f25d487579"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.665751 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.737998 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.753843 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts\") pod \"8600373d-f466-43d4-92fc-1fa938f6e91b\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.753986 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq\") pod \"8600373d-f466-43d4-92fc-1fa938f6e91b\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.754501 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8600373d-f466-43d4-92fc-1fa938f6e91b" (UID: "8600373d-f466-43d4-92fc-1fa938f6e91b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.758463 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq" (OuterVolumeSpecName: "kube-api-access-q7phq") pod "8600373d-f466-43d4-92fc-1fa938f6e91b" (UID: "8600373d-f466-43d4-92fc-1fa938f6e91b"). InnerVolumeSpecName "kube-api-access-q7phq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.855124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4p9x\" (UniqueName: \"kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x\") pod \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.855176 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts\") pod \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.855672 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe6e7565-793b-4555-b6d5-758f87a5b9c7" (UID: "fe6e7565-793b-4555-b6d5-758f87a5b9c7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.856190 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.856217 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.856228 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.857655 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x" (OuterVolumeSpecName: "kube-api-access-w4p9x") pod "fe6e7565-793b-4555-b6d5-758f87a5b9c7" (UID: "fe6e7565-793b-4555-b6d5-758f87a5b9c7"). InnerVolumeSpecName "kube-api-access-w4p9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.959380 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4p9x\" (UniqueName: \"kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.218612 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.218618 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kk8kb" event={"ID":"8600373d-f466-43d4-92fc-1fa938f6e91b","Type":"ContainerDied","Data":"e76ccb54eba21f586b5ec6a78b2a5550037daa9208d52611b08df26c0e3ac56b"} Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.218714 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e76ccb54eba21f586b5ec6a78b2a5550037daa9208d52611b08df26c0e3ac56b" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.220180 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2619-account-create-update-9csh2" event={"ID":"fe6e7565-793b-4555-b6d5-758f87a5b9c7","Type":"ContainerDied","Data":"8526f4c3098ff09c46e3c8957568e817f203fa1293f2cd973998f5c043bb31d2"} Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.220212 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8526f4c3098ff09c46e3c8957568e817f203fa1293f2cd973998f5c043bb31d2" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.220375 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.720140 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.759419 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.759750 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-qzfrf" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="dnsmasq-dns" containerID="cri-o://8a0b04dc02e5d8c871ec744b980efd9b4e195f4024e4ef8187ce38c2f8284265" gracePeriod=10 Jan 21 11:26:16 crc kubenswrapper[4824]: I0121 11:26:16.073664 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:26:16 crc kubenswrapper[4824]: I0121 11:26:16.073889 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:26:16 crc kubenswrapper[4824]: I0121 11:26:16.227770 4824 generic.go:334] "Generic (PLEG): container finished" podID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerID="8a0b04dc02e5d8c871ec744b980efd9b4e195f4024e4ef8187ce38c2f8284265" exitCode=0 Jan 21 11:26:16 crc kubenswrapper[4824]: I0121 11:26:16.227812 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qzfrf" event={"ID":"e94778b1-d36a-47e9-b201-9b7cd1cdb640","Type":"ContainerDied","Data":"8a0b04dc02e5d8c871ec744b980efd9b4e195f4024e4ef8187ce38c2f8284265"} Jan 21 11:26:16 crc kubenswrapper[4824]: I0121 11:26:16.695786 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qzfrf" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.404801 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.420303 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.423307 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.442194 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.449329 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.450146 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504642 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6876j\" (UniqueName: \"kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j\") pod \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504675 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq7qd\" (UniqueName: \"kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd\") pod \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts\") pod \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqhbb\" (UniqueName: \"kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb\") pod \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504795 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts\") pod \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504818 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts\") pod \"5c0ddc66-7696-4816-8e96-340aeb39d242\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504844 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nfhq\" (UniqueName: \"kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq\") pod \"9be37b35-7ad5-434a-8d16-c2f9f1661821\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504875 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts\") pod \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504890 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts\") pod \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504911 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbmg7\" (UniqueName: \"kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7\") pod \"5c0ddc66-7696-4816-8e96-340aeb39d242\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504942 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts\") pod \"9be37b35-7ad5-434a-8d16-c2f9f1661821\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.505007 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvj2q\" (UniqueName: \"kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q\") pod \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.506773 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" (UID: "93c8ce3d-a4f0-4cc7-8592-61653eb09d3c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.506782 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a10a8f17-c27b-4544-8f9d-a6afd9991a8e" (UID: "a10a8f17-c27b-4544-8f9d-a6afd9991a8e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.507240 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b4773e7a-a50c-42e6-bb27-a25b0055b8e3" (UID: "b4773e7a-a50c-42e6-bb27-a25b0055b8e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.507286 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c0ddc66-7696-4816-8e96-340aeb39d242" (UID: "5c0ddc66-7696-4816-8e96-340aeb39d242"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.507490 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9be37b35-7ad5-434a-8d16-c2f9f1661821" (UID: "9be37b35-7ad5-434a-8d16-c2f9f1661821"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.508175 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7b747df8-e61f-4689-a08c-0dc6688f1a0e" (UID: "7b747df8-e61f-4689-a08c-0dc6688f1a0e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.508449 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq" (OuterVolumeSpecName: "kube-api-access-5nfhq") pod "9be37b35-7ad5-434a-8d16-c2f9f1661821" (UID: "9be37b35-7ad5-434a-8d16-c2f9f1661821"). InnerVolumeSpecName "kube-api-access-5nfhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.508983 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd" (OuterVolumeSpecName: "kube-api-access-pq7qd") pod "93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" (UID: "93c8ce3d-a4f0-4cc7-8592-61653eb09d3c"). InnerVolumeSpecName "kube-api-access-pq7qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.515495 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j" (OuterVolumeSpecName: "kube-api-access-6876j") pod "7b747df8-e61f-4689-a08c-0dc6688f1a0e" (UID: "7b747df8-e61f-4689-a08c-0dc6688f1a0e"). InnerVolumeSpecName "kube-api-access-6876j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.523204 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb" (OuterVolumeSpecName: "kube-api-access-pqhbb") pod "a10a8f17-c27b-4544-8f9d-a6afd9991a8e" (UID: "a10a8f17-c27b-4544-8f9d-a6afd9991a8e"). InnerVolumeSpecName "kube-api-access-pqhbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.523257 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7" (OuterVolumeSpecName: "kube-api-access-lbmg7") pod "5c0ddc66-7696-4816-8e96-340aeb39d242" (UID: "5c0ddc66-7696-4816-8e96-340aeb39d242"). InnerVolumeSpecName "kube-api-access-lbmg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.526664 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q" (OuterVolumeSpecName: "kube-api-access-wvj2q") pod "b4773e7a-a50c-42e6-bb27-a25b0055b8e3" (UID: "b4773e7a-a50c-42e6-bb27-a25b0055b8e3"). InnerVolumeSpecName "kube-api-access-wvj2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607544 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6876j\" (UniqueName: \"kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607805 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq7qd\" (UniqueName: \"kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607815 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607824 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqhbb\" (UniqueName: \"kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607832 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607841 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607848 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nfhq\" (UniqueName: \"kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607858 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607865 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607873 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbmg7\" (UniqueName: \"kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607881 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607890 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvj2q\" (UniqueName: \"kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.705343 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.810624 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb\") pod \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.810704 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb\") pod \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.810755 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc\") pod \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.810786 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config\") pod \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.810848 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9jhb\" (UniqueName: \"kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb\") pod \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.814663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb" (OuterVolumeSpecName: "kube-api-access-q9jhb") pod "e94778b1-d36a-47e9-b201-9b7cd1cdb640" (UID: "e94778b1-d36a-47e9-b201-9b7cd1cdb640"). InnerVolumeSpecName "kube-api-access-q9jhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.837174 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e94778b1-d36a-47e9-b201-9b7cd1cdb640" (UID: "e94778b1-d36a-47e9-b201-9b7cd1cdb640"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.839031 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e94778b1-d36a-47e9-b201-9b7cd1cdb640" (UID: "e94778b1-d36a-47e9-b201-9b7cd1cdb640"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.839373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config" (OuterVolumeSpecName: "config") pod "e94778b1-d36a-47e9-b201-9b7cd1cdb640" (UID: "e94778b1-d36a-47e9-b201-9b7cd1cdb640"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.839976 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e94778b1-d36a-47e9-b201-9b7cd1cdb640" (UID: "e94778b1-d36a-47e9-b201-9b7cd1cdb640"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.913719 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9jhb\" (UniqueName: \"kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.913741 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.913750 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.913759 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.913767 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.242182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3c9f-account-create-update-rtprr" event={"ID":"5c0ddc66-7696-4816-8e96-340aeb39d242","Type":"ContainerDied","Data":"e9e93c5745e7fe74ef1e643164e9a0a4e165267deeb109faf4df81c36f336691"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.242223 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9e93c5745e7fe74ef1e643164e9a0a4e165267deeb109faf4df81c36f336691" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.242290 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.243866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ngtnj" event={"ID":"9be37b35-7ad5-434a-8d16-c2f9f1661821","Type":"ContainerDied","Data":"e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.243888 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.243936 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.245148 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-twgjw" event={"ID":"7b747df8-e61f-4689-a08c-0dc6688f1a0e","Type":"ContainerDied","Data":"5cbfd6e560404567a0bbc9ae88f4245a4c606e0f65ba74f1fdb8d8f25d487579"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.245172 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cbfd6e560404567a0bbc9ae88f4245a4c606e0f65ba74f1fdb8d8f25d487579" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.245213 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.246546 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qzfrf" event={"ID":"e94778b1-d36a-47e9-b201-9b7cd1cdb640","Type":"ContainerDied","Data":"48fb0631b23867cac173ec7a9df005de8edd1cb0df987f8dceb763f2434dd5b7"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.246575 4824 scope.go:117] "RemoveContainer" containerID="8a0b04dc02e5d8c871ec744b980efd9b4e195f4024e4ef8187ce38c2f8284265" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.246671 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.248845 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-84f4b" event={"ID":"5240d35b-cbf1-472d-91da-debf418dd208","Type":"ContainerStarted","Data":"3d07104d28aceb5ad31acec840c18d559f5c98c4390845b748361ff21f72d204"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.253663 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0662-account-create-update-j7bhp" event={"ID":"b4773e7a-a50c-42e6-bb27-a25b0055b8e3","Type":"ContainerDied","Data":"6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.253695 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.254718 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.261068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nskxc" event={"ID":"a10a8f17-c27b-4544-8f9d-a6afd9991a8e","Type":"ContainerDied","Data":"d2e64d054f528de025088752cbba2a64296b9807a6337f5fdd775173aefc9828"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.261097 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2e64d054f528de025088752cbba2a64296b9807a6337f5fdd775173aefc9828" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.261076 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.265575 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-60f6-account-create-update-pjh6n" event={"ID":"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c","Type":"ContainerDied","Data":"a60e0d515d078f64392d5d1ca5e1034d522ce6ad7bf98f9a82fa5e036ed775d2"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.265609 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a60e0d515d078f64392d5d1ca5e1034d522ce6ad7bf98f9a82fa5e036ed775d2" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.265618 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.271392 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-84f4b" podStartSLOduration=2.9813106400000002 podStartE2EDuration="7.271381381s" podCreationTimestamp="2026-01-21 11:26:11 +0000 UTC" firstStartedPulling="2026-01-21 11:26:12.989679987 +0000 UTC m=+915.282709279" lastFinishedPulling="2026-01-21 11:26:17.279750729 +0000 UTC m=+919.572780020" observedRunningTime="2026-01-21 11:26:18.265777851 +0000 UTC m=+920.558807143" watchObservedRunningTime="2026-01-21 11:26:18.271381381 +0000 UTC m=+920.564410673" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.296160 4824 scope.go:117] "RemoveContainer" containerID="a517f23f781efca10654bdff86d5e2efc32071ea09d5af5829095603d962c629" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.342664 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.347720 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:26:19 crc kubenswrapper[4824]: I0121 11:26:19.273749 4824 generic.go:334] "Generic (PLEG): container finished" podID="5240d35b-cbf1-472d-91da-debf418dd208" containerID="3d07104d28aceb5ad31acec840c18d559f5c98c4390845b748361ff21f72d204" exitCode=0 Jan 21 11:26:19 crc kubenswrapper[4824]: I0121 11:26:19.273784 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-84f4b" event={"ID":"5240d35b-cbf1-472d-91da-debf418dd208","Type":"ContainerDied","Data":"3d07104d28aceb5ad31acec840c18d559f5c98c4390845b748361ff21f72d204"} Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.056175 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" path="/var/lib/kubelet/pods/e94778b1-d36a-47e9-b201-9b7cd1cdb640/volumes" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.553663 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.652827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data\") pod \"5240d35b-cbf1-472d-91da-debf418dd208\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.652902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4nqn\" (UniqueName: \"kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn\") pod \"5240d35b-cbf1-472d-91da-debf418dd208\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.652954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle\") pod \"5240d35b-cbf1-472d-91da-debf418dd208\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.658900 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn" (OuterVolumeSpecName: "kube-api-access-f4nqn") pod "5240d35b-cbf1-472d-91da-debf418dd208" (UID: "5240d35b-cbf1-472d-91da-debf418dd208"). InnerVolumeSpecName "kube-api-access-f4nqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.671162 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5240d35b-cbf1-472d-91da-debf418dd208" (UID: "5240d35b-cbf1-472d-91da-debf418dd208"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.683663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data" (OuterVolumeSpecName: "config-data") pod "5240d35b-cbf1-472d-91da-debf418dd208" (UID: "5240d35b-cbf1-472d-91da-debf418dd208"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.754443 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.754468 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4nqn\" (UniqueName: \"kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.754477 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.279995 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280638 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="dnsmasq-dns" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280656 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="dnsmasq-dns" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280698 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280705 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280714 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c0ddc66-7696-4816-8e96-340aeb39d242" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280719 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c0ddc66-7696-4816-8e96-340aeb39d242" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280730 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a10a8f17-c27b-4544-8f9d-a6afd9991a8e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280736 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a10a8f17-c27b-4544-8f9d-a6afd9991a8e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280763 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8600373d-f466-43d4-92fc-1fa938f6e91b" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280769 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8600373d-f466-43d4-92fc-1fa938f6e91b" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280775 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b747df8-e61f-4689-a08c-0dc6688f1a0e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280782 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b747df8-e61f-4689-a08c-0dc6688f1a0e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280794 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5240d35b-cbf1-472d-91da-debf418dd208" containerName="keystone-db-sync" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280800 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5240d35b-cbf1-472d-91da-debf418dd208" containerName="keystone-db-sync" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280818 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4773e7a-a50c-42e6-bb27-a25b0055b8e3" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280840 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4773e7a-a50c-42e6-bb27-a25b0055b8e3" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280847 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="init" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280852 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="init" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280858 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9be37b35-7ad5-434a-8d16-c2f9f1661821" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280863 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9be37b35-7ad5-434a-8d16-c2f9f1661821" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280876 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe6e7565-793b-4555-b6d5-758f87a5b9c7" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280881 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe6e7565-793b-4555-b6d5-758f87a5b9c7" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281072 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281084 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8600373d-f466-43d4-92fc-1fa938f6e91b" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281093 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a10a8f17-c27b-4544-8f9d-a6afd9991a8e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281102 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9be37b35-7ad5-434a-8d16-c2f9f1661821" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281109 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b747df8-e61f-4689-a08c-0dc6688f1a0e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281136 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5240d35b-cbf1-472d-91da-debf418dd208" containerName="keystone-db-sync" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281143 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe6e7565-793b-4555-b6d5-758f87a5b9c7" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281155 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c0ddc66-7696-4816-8e96-340aeb39d242" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281164 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="dnsmasq-dns" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281173 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4773e7a-a50c-42e6-bb27-a25b0055b8e3" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.285596 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.296897 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-84f4b" event={"ID":"5240d35b-cbf1-472d-91da-debf418dd208","Type":"ContainerDied","Data":"9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581"} Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.296939 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.297071 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.313557 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.342801 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sw5v2"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.343700 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.353723 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.353817 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.353969 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.354078 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.354588 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxhzv" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.363877 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.363923 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.363993 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5zvt\" (UniqueName: \"kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.364016 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.364063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.364102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.369056 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sw5v2"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.444108 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-jrsn8"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.445022 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.446698 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-cj2g7" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.447052 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.455183 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-jrsn8"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465264 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465323 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztr5x\" (UniqueName: \"kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465453 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465518 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5zvt\" (UniqueName: \"kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465671 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465791 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.466316 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.466516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.467005 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.467549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.467602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.475556 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.478801 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.480201 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.480570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.485724 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5zvt\" (UniqueName: \"kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.491687 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.567320 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgmtk\" (UniqueName: \"kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrdwp\" (UniqueName: \"kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568138 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568223 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568287 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568466 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568531 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568612 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568710 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568845 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.569006 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.569075 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.569145 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.569223 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztr5x\" (UniqueName: \"kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.574230 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.574747 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.575086 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.575274 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.580639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.590315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztr5x\" (UniqueName: \"kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.601253 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.619750 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.631900 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.634546 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.641290 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.663772 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.665461 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6vgh6"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.666384 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.676929 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrdwp\" (UniqueName: \"kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677010 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677053 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677089 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677119 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677297 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677360 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8x9b\" (UniqueName: \"kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgmtk\" (UniqueName: \"kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.679424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.680009 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.680343 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n8hrk" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.680539 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.680870 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.683769 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.684387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.684734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.685535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.686347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.690827 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.698585 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrdwp\" (UniqueName: \"kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.699664 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgmtk\" (UniqueName: \"kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.702535 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6vgh6"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.762740 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778290 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778333 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkb9f\" (UniqueName: \"kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778365 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778390 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778456 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778510 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8x9b\" (UniqueName: \"kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.779520 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.779536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.779717 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.779796 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.780319 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.794886 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8x9b\" (UniqueName: \"kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.823235 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.880324 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.880646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.880989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkb9f\" (UniqueName: \"kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.881040 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.881453 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.881507 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.884140 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.884344 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.885019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.896298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkb9f\" (UniqueName: \"kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.009255 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.017508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.118671 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:22 crc kubenswrapper[4824]: W0121 11:26:22.143195 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcfd76c3_d71d_4f02_af00_d01b113c4d2e.slice/crio-158d502ef9a1830d418ad9c0bd801b29d56915eafcb443ee997a5f301edbf91f WatchSource:0}: Error finding container 158d502ef9a1830d418ad9c0bd801b29d56915eafcb443ee997a5f301edbf91f: Status 404 returned error can't find the container with id 158d502ef9a1830d418ad9c0bd801b29d56915eafcb443ee997a5f301edbf91f Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.171726 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sw5v2"] Jan 21 11:26:22 crc kubenswrapper[4824]: W0121 11:26:22.178132 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f0bb2ed_9f0d_4d49_b224_8d05b5de0829.slice/crio-0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853 WatchSource:0}: Error finding container 0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853: Status 404 returned error can't find the container with id 0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853 Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.245792 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-jrsn8"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.374122 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.379233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" event={"ID":"bcfd76c3-d71d-4f02-af00-d01b113c4d2e","Type":"ContainerStarted","Data":"158d502ef9a1830d418ad9c0bd801b29d56915eafcb443ee997a5f301edbf91f"} Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.382868 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jrsn8" event={"ID":"cadf2d38-24f9-43ab-8e29-803d97a0ff06","Type":"ContainerStarted","Data":"f8ce50a0aaa511cfcd5c91e987379f42c2650248768a77a65a4355ed457a06c8"} Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.386078 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sw5v2" event={"ID":"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829","Type":"ContainerStarted","Data":"0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853"} Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.404934 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.406377 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.409239 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4bzft" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.409435 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.409657 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.409787 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.417677 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.432474 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ngq69"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.433753 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.439498 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8q76f" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.439751 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.490076 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvbdv\" (UniqueName: \"kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508755 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508926 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flpqh\" (UniqueName: \"kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.509024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.509189 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.509239 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.509282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.509318 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.526470 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.531084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.531264 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.541245 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ngq69"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.563211 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.573553 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-rlp2t"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.574444 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.576384 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.576906 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2xjd5" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.577035 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.583814 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rlp2t"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.591104 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6vgh6"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.610294 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.610502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.610637 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.610782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.610922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.611550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvbdv\" (UniqueName: \"kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.611700 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.611793 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msmnx\" (UniqueName: \"kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.611875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.611971 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612161 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612497 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4wc2\" (UniqueName: \"kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612741 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flpqh\" (UniqueName: \"kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612823 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612915 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613021 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613178 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612934 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613318 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613653 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.615578 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.617308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.618200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.618437 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.618514 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.623728 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.625561 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvbdv\" (UniqueName: \"kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.632974 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.633386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flpqh\" (UniqueName: \"kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.635583 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msmnx\" (UniqueName: \"kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715287 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715318 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715351 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4wc2\" (UniqueName: \"kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715401 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715425 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715464 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715498 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715532 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715583 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715649 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.716035 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.716229 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.716289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.721422 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.721619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.721849 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.722617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.722704 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.724142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.725291 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.727574 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.734076 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msmnx\" (UniqueName: \"kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.734619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4wc2\" (UniqueName: \"kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.749511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.754712 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-t2txg"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.756000 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.758324 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.758374 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.758387 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-49xnt" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.762481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.773737 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-t2txg"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.780583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.816838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.817437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpjkl\" (UniqueName: \"kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.817466 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.850298 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.919435 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpjkl\" (UniqueName: \"kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.919473 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.919535 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.925519 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.929570 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.936794 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpjkl\" (UniqueName: \"kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.021724 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.043666 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.070049 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.080422 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.166434 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.193758 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rlp2t"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.335632 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:23 crc kubenswrapper[4824]: W0121 11:26:23.372019 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod262733b9_4179_49e3_aee9_b62197cc89ba.slice/crio-2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf WatchSource:0}: Error finding container 2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf: Status 404 returned error can't find the container with id 2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.375484 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ngq69"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.405166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rlp2t" event={"ID":"0a6c0628-4522-4bb7-8a82-cc2e019eca2a","Type":"ContainerStarted","Data":"877238fc09eb99e400629467da99c2ba4a6e53c0afc16677b1308d3c52849ae9"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.409442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sw5v2" event={"ID":"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829","Type":"ContainerStarted","Data":"68285af994e4fe82e817010ea6bff330d5655f091788a58ca5b521d0e79564f1"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.411134 4824 generic.go:334] "Generic (PLEG): container finished" podID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerID="7af315bc544e0ed568b3db042720a7c28e2f59b1fb260d2208c816705728d7f5" exitCode=0 Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.411192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" event={"ID":"8214df67-a984-4a5b-bcbb-6b16ab42f5fe","Type":"ContainerDied","Data":"7af315bc544e0ed568b3db042720a7c28e2f59b1fb260d2208c816705728d7f5"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.411216 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" event={"ID":"8214df67-a984-4a5b-bcbb-6b16ab42f5fe","Type":"ContainerStarted","Data":"a7ed7bee402dc3db5565fe550b060e37c1771f55a94528a7a39dbd2741c1ec27"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.415321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6vgh6" event={"ID":"08b8c363-a8c1-4dfe-8fd1-1af2d055a305","Type":"ContainerStarted","Data":"97721ccfe8abdf2324794ac67debd67a8fc655d3d7ab538bf2cb4b35e35f8386"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.418241 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerStarted","Data":"9fa11b86f012b31aa61fb0528e706362f80104db47ab25f797420eb04f92a3fa"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.420215 4824 generic.go:334] "Generic (PLEG): container finished" podID="bcfd76c3-d71d-4f02-af00-d01b113c4d2e" containerID="27ee2d89193499812ebe436242e9b3d7756adae2afdbb8a226ae95bcc2f9a4f1" exitCode=0 Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.420824 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" event={"ID":"bcfd76c3-d71d-4f02-af00-d01b113c4d2e","Type":"ContainerDied","Data":"27ee2d89193499812ebe436242e9b3d7756adae2afdbb8a226ae95bcc2f9a4f1"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.423555 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ngq69" event={"ID":"262733b9-4179-49e3-aee9-b62197cc89ba","Type":"ContainerStarted","Data":"2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.429631 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sw5v2" podStartSLOduration=2.429616409 podStartE2EDuration="2.429616409s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:23.423702974 +0000 UTC m=+925.716732267" watchObservedRunningTime="2026-01-21 11:26:23.429616409 +0000 UTC m=+925.722645701" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.432058 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerStarted","Data":"cb87ef02c610bd346ad76ef8c74796d1ec1bf79d6d471b5705c3e5bb8a69ec3b"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.650670 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.657807 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-t2txg"] Jan 21 11:26:23 crc kubenswrapper[4824]: W0121 11:26:23.665795 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d0e8430_4bfe_4e37_8c22_502dd5444af4.slice/crio-51a42c61ed30de7289f1aa596c3516c76d7926804f43324d5132c6f93e7e5c66 WatchSource:0}: Error finding container 51a42c61ed30de7289f1aa596c3516c76d7926804f43324d5132c6f93e7e5c66: Status 404 returned error can't find the container with id 51a42c61ed30de7289f1aa596c3516c76d7926804f43324d5132c6f93e7e5c66 Jan 21 11:26:23 crc kubenswrapper[4824]: W0121 11:26:23.684757 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda19b6730_b7aa_4319_8127_c32d7874a471.slice/crio-936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209 WatchSource:0}: Error finding container 936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209: Status 404 returned error can't find the container with id 936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209 Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.741668 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844568 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5zvt\" (UniqueName: \"kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844652 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844868 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844936 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844997 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.853761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt" (OuterVolumeSpecName: "kube-api-access-h5zvt") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "kube-api-access-h5zvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.864619 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.868313 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config" (OuterVolumeSpecName: "config") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.868487 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.871602 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.878639 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947196 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5zvt\" (UniqueName: \"kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947221 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947231 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947239 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947247 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947254 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.451600 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerStarted","Data":"b624461c7fd817bff45a1813fc59b08421416d9686db5ebe66efe4049b30534b"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.454417 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerStarted","Data":"859f2e0dceb5ba0b5dc20dc8542580aaaceec1b327c7c16135166be30bdad2a1"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.454453 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerStarted","Data":"51a42c61ed30de7289f1aa596c3516c76d7926804f43324d5132c6f93e7e5c66"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.458200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" event={"ID":"8214df67-a984-4a5b-bcbb-6b16ab42f5fe","Type":"ContainerStarted","Data":"56442b913a4db71125dc75408dc50b059e8519640a2cc312a0c8277e0b1aefab"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.458995 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.461309 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t2txg" event={"ID":"a19b6730-b7aa-4319-8127-c32d7874a471","Type":"ContainerStarted","Data":"49abf3af76b6411b5212e8cfa13e8c4145a6700e1d310d3281c5d847c1a1f457"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.461337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t2txg" event={"ID":"a19b6730-b7aa-4319-8127-c32d7874a471","Type":"ContainerStarted","Data":"936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.475759 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" podStartSLOduration=3.475748282 podStartE2EDuration="3.475748282s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:24.475139695 +0000 UTC m=+926.768168987" watchObservedRunningTime="2026-01-21 11:26:24.475748282 +0000 UTC m=+926.768777574" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.484034 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.484064 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" event={"ID":"bcfd76c3-d71d-4f02-af00-d01b113c4d2e","Type":"ContainerDied","Data":"158d502ef9a1830d418ad9c0bd801b29d56915eafcb443ee997a5f301edbf91f"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.484089 4824 scope.go:117] "RemoveContainer" containerID="27ee2d89193499812ebe436242e9b3d7756adae2afdbb8a226ae95bcc2f9a4f1" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.491021 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-t2txg" podStartSLOduration=2.491005481 podStartE2EDuration="2.491005481s" podCreationTimestamp="2026-01-21 11:26:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:24.485381483 +0000 UTC m=+926.778410775" watchObservedRunningTime="2026-01-21 11:26:24.491005481 +0000 UTC m=+926.784034773" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.624024 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.633308 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.500370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerStarted","Data":"7f3d59f931487b81345a89a9579c3ee29598f0123e3fc5337b9d33bd8702f80f"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.500455 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-log" containerID="cri-o://b624461c7fd817bff45a1813fc59b08421416d9686db5ebe66efe4049b30534b" gracePeriod=30 Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.500517 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-httpd" containerID="cri-o://7f3d59f931487b81345a89a9579c3ee29598f0123e3fc5337b9d33bd8702f80f" gracePeriod=30 Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.510900 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerStarted","Data":"0f7acbd726575b92b89853489927b20157e8c557e9675f9935d803c1e04cda43"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.510969 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-log" containerID="cri-o://859f2e0dceb5ba0b5dc20dc8542580aaaceec1b327c7c16135166be30bdad2a1" gracePeriod=30 Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.510986 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-httpd" containerID="cri-o://0f7acbd726575b92b89853489927b20157e8c557e9675f9935d803c1e04cda43" gracePeriod=30 Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.540258 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.540241329 podStartE2EDuration="4.540241329s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:25.539648792 +0000 UTC m=+927.832678084" watchObservedRunningTime="2026-01-21 11:26:25.540241329 +0000 UTC m=+927.833270621" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.542642 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.542629685 podStartE2EDuration="4.542629685s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:25.521819491 +0000 UTC m=+927.814848783" watchObservedRunningTime="2026-01-21 11:26:25.542629685 +0000 UTC m=+927.835658976" Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.073537 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcfd76c3-d71d-4f02-af00-d01b113c4d2e" path="/var/lib/kubelet/pods/bcfd76c3-d71d-4f02-af00-d01b113c4d2e/volumes" Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.524558 4824 generic.go:334] "Generic (PLEG): container finished" podID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerID="7f3d59f931487b81345a89a9579c3ee29598f0123e3fc5337b9d33bd8702f80f" exitCode=0 Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.524772 4824 generic.go:334] "Generic (PLEG): container finished" podID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerID="b624461c7fd817bff45a1813fc59b08421416d9686db5ebe66efe4049b30534b" exitCode=143 Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.524635 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerDied","Data":"7f3d59f931487b81345a89a9579c3ee29598f0123e3fc5337b9d33bd8702f80f"} Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.524837 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerDied","Data":"b624461c7fd817bff45a1813fc59b08421416d9686db5ebe66efe4049b30534b"} Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.526324 4824 generic.go:334] "Generic (PLEG): container finished" podID="4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" containerID="68285af994e4fe82e817010ea6bff330d5655f091788a58ca5b521d0e79564f1" exitCode=0 Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.526389 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sw5v2" event={"ID":"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829","Type":"ContainerDied","Data":"68285af994e4fe82e817010ea6bff330d5655f091788a58ca5b521d0e79564f1"} Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.530569 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerID="0f7acbd726575b92b89853489927b20157e8c557e9675f9935d803c1e04cda43" exitCode=0 Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.530610 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerID="859f2e0dceb5ba0b5dc20dc8542580aaaceec1b327c7c16135166be30bdad2a1" exitCode=143 Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.531514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerDied","Data":"0f7acbd726575b92b89853489927b20157e8c557e9675f9935d803c1e04cda43"} Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.531557 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerDied","Data":"859f2e0dceb5ba0b5dc20dc8542580aaaceec1b327c7c16135166be30bdad2a1"} Jan 21 11:26:30 crc kubenswrapper[4824]: I0121 11:26:30.574221 4824 generic.go:334] "Generic (PLEG): container finished" podID="a19b6730-b7aa-4319-8127-c32d7874a471" containerID="49abf3af76b6411b5212e8cfa13e8c4145a6700e1d310d3281c5d847c1a1f457" exitCode=0 Jan 21 11:26:30 crc kubenswrapper[4824]: I0121 11:26:30.574256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t2txg" event={"ID":"a19b6730-b7aa-4319-8127-c32d7874a471","Type":"ContainerDied","Data":"49abf3af76b6411b5212e8cfa13e8c4145a6700e1d310d3281c5d847c1a1f457"} Jan 21 11:26:32 crc kubenswrapper[4824]: I0121 11:26:32.011075 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:32 crc kubenswrapper[4824]: I0121 11:26:32.066527 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:32 crc kubenswrapper[4824]: I0121 11:26:32.066732 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" containerID="cri-o://862050906c3faeb783c3cf1a55874b97964bf3e73dcab5972c6e49f1ea08c530" gracePeriod=10 Jan 21 11:26:32 crc kubenswrapper[4824]: I0121 11:26:32.593378 4824 generic.go:334] "Generic (PLEG): container finished" podID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerID="862050906c3faeb783c3cf1a55874b97964bf3e73dcab5972c6e49f1ea08c530" exitCode=0 Jan 21 11:26:32 crc kubenswrapper[4824]: I0121 11:26:32.593408 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" event={"ID":"c4a81903-5e1f-4d9c-bacf-60639069db18","Type":"ContainerDied","Data":"862050906c3faeb783c3cf1a55874b97964bf3e73dcab5972c6e49f1ea08c530"} Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.498094 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531564 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531607 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztr5x\" (UniqueName: \"kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531683 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531731 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.542466 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.542495 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.542570 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts" (OuterVolumeSpecName: "scripts") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.542619 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x" (OuterVolumeSpecName: "kube-api-access-ztr5x") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "kube-api-access-ztr5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.550203 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.551130 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data" (OuterVolumeSpecName: "config-data") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.607734 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sw5v2" event={"ID":"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829","Type":"ContainerDied","Data":"0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853"} Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.607769 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.607776 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633036 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633062 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633071 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztr5x\" (UniqueName: \"kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633079 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633088 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633097 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.563748 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sw5v2"] Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.569012 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sw5v2"] Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.665191 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5l2gg"] Jan 21 11:26:35 crc kubenswrapper[4824]: E0121 11:26:35.665885 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcfd76c3-d71d-4f02-af00-d01b113c4d2e" containerName="init" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.665904 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcfd76c3-d71d-4f02-af00-d01b113c4d2e" containerName="init" Jan 21 11:26:35 crc kubenswrapper[4824]: E0121 11:26:35.665950 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" containerName="keystone-bootstrap" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.665976 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" containerName="keystone-bootstrap" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.666191 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" containerName="keystone-bootstrap" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.666225 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcfd76c3-d71d-4f02-af00-d01b113c4d2e" containerName="init" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.666876 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.668411 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.668853 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxhzv" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.668914 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.669007 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.669738 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.671981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5l2gg"] Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.719065 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748515 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lj6r\" (UniqueName: \"kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748647 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748817 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.849994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.850064 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lj6r\" (UniqueName: \"kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.850098 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.850130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.850144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.850210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.853402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.853461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.853944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.854254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.855039 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.863272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lj6r\" (UniqueName: \"kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.985949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:36 crc kubenswrapper[4824]: I0121 11:26:36.060377 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" path="/var/lib/kubelet/pods/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829/volumes" Jan 21 11:26:39 crc kubenswrapper[4824]: I0121 11:26:39.935992 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:39 crc kubenswrapper[4824]: I0121 11:26:39.954283 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:39 crc kubenswrapper[4824]: I0121 11:26:39.957372 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.105850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.105893 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpjkl\" (UniqueName: \"kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl\") pod \"a19b6730-b7aa-4319-8127-c32d7874a471\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.105911 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.105949 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4wc2\" (UniqueName: \"kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.105987 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle\") pod \"a19b6730-b7aa-4319-8127-c32d7874a471\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106017 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106034 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106058 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106095 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106118 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106157 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config\") pod \"a19b6730-b7aa-4319-8127-c32d7874a471\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106177 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106645 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvbdv\" (UniqueName: \"kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106733 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106771 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106776 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs" (OuterVolumeSpecName: "logs") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106792 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106811 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs" (OuterVolumeSpecName: "logs") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106824 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.107142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.107369 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.108445 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.108463 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.108473 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.108480 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.110665 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv" (OuterVolumeSpecName: "kube-api-access-mvbdv") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "kube-api-access-mvbdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.112439 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.112543 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2" (OuterVolumeSpecName: "kube-api-access-f4wc2") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "kube-api-access-f4wc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.112836 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.115250 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts" (OuterVolumeSpecName: "scripts") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.115323 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl" (OuterVolumeSpecName: "kube-api-access-dpjkl") pod "a19b6730-b7aa-4319-8127-c32d7874a471" (UID: "a19b6730-b7aa-4319-8127-c32d7874a471"). InnerVolumeSpecName "kube-api-access-dpjkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.115256 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts" (OuterVolumeSpecName: "scripts") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.132211 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.134905 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.143106 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a19b6730-b7aa-4319-8127-c32d7874a471" (UID: "a19b6730-b7aa-4319-8127-c32d7874a471"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.145896 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data" (OuterVolumeSpecName: "config-data") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.146316 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config" (OuterVolumeSpecName: "config") pod "a19b6730-b7aa-4319-8127-c32d7874a471" (UID: "a19b6730-b7aa-4319-8127-c32d7874a471"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.149099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.153915 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.155443 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data" (OuterVolumeSpecName: "config-data") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210391 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210417 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210426 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210434 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210442 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpjkl\" (UniqueName: \"kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210452 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210460 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4wc2\" (UniqueName: \"kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210467 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210494 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210502 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210514 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210521 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210529 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210536 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210543 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvbdv\" (UniqueName: \"kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.226442 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.231270 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.311672 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.311712 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.321123 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.321249 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-flpqh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-ngq69_openstack(262733b9-4179-49e3-aee9-b62197cc89ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.322760 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-ngq69" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.644899 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerDied","Data":"51a42c61ed30de7289f1aa596c3516c76d7926804f43324d5132c6f93e7e5c66"} Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.644921 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.644945 4824 scope.go:117] "RemoveContainer" containerID="0f7acbd726575b92b89853489927b20157e8c557e9675f9935d803c1e04cda43" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.646254 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t2txg" event={"ID":"a19b6730-b7aa-4319-8127-c32d7874a471","Type":"ContainerDied","Data":"936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209"} Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.646275 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.646334 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.653071 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.653870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerDied","Data":"cb87ef02c610bd346ad76ef8c74796d1ec1bf79d6d471b5705c3e5bb8a69ec3b"} Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.654420 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-ngq69" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.689970 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.699905 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.707332 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.712884 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.722683 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.723064 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a19b6730-b7aa-4319-8127-c32d7874a471" containerName="neutron-db-sync" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723078 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a19b6730-b7aa-4319-8127-c32d7874a471" containerName="neutron-db-sync" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.723092 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723097 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.723113 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723118 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.723130 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723135 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.723153 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723158 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723364 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723378 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723387 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723398 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a19b6730-b7aa-4319-8127-c32d7874a471" containerName="neutron-db-sync" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723405 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.727155 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.728200 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.729216 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.732511 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.732743 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.732881 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.733024 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4bzft" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.733387 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.733474 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.739767 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.746390 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820057 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820111 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820252 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820275 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820389 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8grh\" (UniqueName: \"kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.921596 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922012 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922164 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qwlh\" (UniqueName: \"kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922198 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922344 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8grh\" (UniqueName: \"kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922388 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922518 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922727 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922787 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922824 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922866 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922921 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922976 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.923002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.923132 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.927599 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.927876 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.928853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.928904 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.935340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8grh\" (UniqueName: \"kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.940946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024013 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024053 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024082 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024115 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024151 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024185 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qwlh\" (UniqueName: \"kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024185 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024206 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.027768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.028337 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.028654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.028811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.037467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qwlh\" (UniqueName: \"kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.041020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.050031 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.063517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.160306 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.166150 4824 scope.go:117] "RemoveContainer" containerID="859f2e0dceb5ba0b5dc20dc8542580aaaceec1b327c7c16135166be30bdad2a1" Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.166150 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.166343 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-msmnx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-rlp2t_openstack(0a6c0628-4522-4bb7-8a82-cc2e019eca2a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.167919 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-rlp2t" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.170009 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.174261 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.201982 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.243831 4824 scope.go:117] "RemoveContainer" containerID="7f3d59f931487b81345a89a9579c3ee29598f0123e3fc5337b9d33bd8702f80f" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.319068 4824 scope.go:117] "RemoveContainer" containerID="b624461c7fd817bff45a1813fc59b08421416d9686db5ebe66efe4049b30534b" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.342665 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.342739 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.342829 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.342918 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.342996 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xgrw\" (UniqueName: \"kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343028 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343362 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343403 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343807 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343834 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clj2h\" (UniqueName: \"kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343872 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.353762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw" (OuterVolumeSpecName: "kube-api-access-9xgrw") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "kube-api-access-9xgrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.399296 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.406873 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config" (OuterVolumeSpecName: "config") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.430364 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.430769 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.430787 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.430800 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="init" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.430805 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="init" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.430984 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.443061 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.446434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.446816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clj2h\" (UniqueName: \"kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.446866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.446946 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447078 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447109 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447562 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xgrw\" (UniqueName: \"kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447582 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447591 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447948 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.448000 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-49xnt" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.448559 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.448755 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.456478 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.464970 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.465644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.467492 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.469900 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.474842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clj2h\" (UniqueName: \"kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.510105 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.513080 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.526859 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.550831 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.550907 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.551922 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.552004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.552032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x8cf\" (UniqueName: \"kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.552112 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.552153 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.552165 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.654654 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.654737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.654888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.654936 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.654972 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x8cf\" (UniqueName: \"kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.660834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.662412 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.662457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.672689 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.680030 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5l2gg"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.688481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x8cf\" (UniqueName: \"kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.694498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jrsn8" event={"ID":"cadf2d38-24f9-43ab-8e29-803d97a0ff06","Type":"ContainerStarted","Data":"5dcf0246f4ab2f2f4f3230d6a47ec6ac1102fcb34b4edb4d89a73c99d792f5ff"} Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.714164 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" event={"ID":"c4a81903-5e1f-4d9c-bacf-60639069db18","Type":"ContainerDied","Data":"19d64cd123fe9ef3646e70cf5808fefa9acba69592569cd21640e9399cbca5d9"} Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.714209 4824 scope.go:117] "RemoveContainer" containerID="862050906c3faeb783c3cf1a55874b97964bf3e73dcab5972c6e49f1ea08c530" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.714327 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.746634 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6vgh6" event={"ID":"08b8c363-a8c1-4dfe-8fd1-1af2d055a305","Type":"ContainerStarted","Data":"5f4a1e6bcd0c138e59cd1e69d47cce9046bae28ab6b9ce57bd7be2ad0c6a4e61"} Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.748592 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-jrsn8" podStartSLOduration=1.8425048130000001 podStartE2EDuration="20.748578322s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="2026-01-21 11:26:22.260048638 +0000 UTC m=+924.553077930" lastFinishedPulling="2026-01-21 11:26:41.166122146 +0000 UTC m=+943.459151439" observedRunningTime="2026-01-21 11:26:41.7104694 +0000 UTC m=+944.003498681" watchObservedRunningTime="2026-01-21 11:26:41.748578322 +0000 UTC m=+944.041607613" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.756938 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.758402 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerStarted","Data":"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c"} Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.764735 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.765666 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-rlp2t" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.765671 4824 scope.go:117] "RemoveContainer" containerID="0aacefa5ca6057f616f2afc320e43a0e95b25311f579f1cc69ade834fb63b2f0" Jan 21 11:26:41 crc kubenswrapper[4824]: W0121 11:26:41.769070 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46f0fa4b_479a_42c5_b9b2_3f30a079be48.slice/crio-5dc758f1e11243d2a910301ab89754d05c98dfa5446bb49c248a5833a44c6ce9 WatchSource:0}: Error finding container 5dc758f1e11243d2a910301ab89754d05c98dfa5446bb49c248a5833a44c6ce9: Status 404 returned error can't find the container with id 5dc758f1e11243d2a910301ab89754d05c98dfa5446bb49c248a5833a44c6ce9 Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.771171 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.779981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.780332 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6vgh6" podStartSLOduration=2.210949144 podStartE2EDuration="20.780315956s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="2026-01-21 11:26:22.566259702 +0000 UTC m=+924.859288994" lastFinishedPulling="2026-01-21 11:26:41.135626515 +0000 UTC m=+943.428655806" observedRunningTime="2026-01-21 11:26:41.762243607 +0000 UTC m=+944.055272899" watchObservedRunningTime="2026-01-21 11:26:41.780315956 +0000 UTC m=+944.073345248" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.848867 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.939206 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.062879 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" path="/var/lib/kubelet/pods/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859/volumes" Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.063692 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" path="/var/lib/kubelet/pods/7d0e8430-4bfe-4e37-8c22-502dd5444af4/volumes" Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.064718 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" path="/var/lib/kubelet/pods/c4a81903-5e1f-4d9c-bacf-60639069db18/volumes" Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.288037 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:26:42 crc kubenswrapper[4824]: W0121 11:26:42.303499 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9302662_76b1_4acc_b9e3_ffde925d34b2.slice/crio-2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d WatchSource:0}: Error finding container 2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d: Status 404 returned error can't find the container with id 2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.767636 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerID="f5715a5c2dbadf161e906011876459ee94ba9a1523d9455c1d55ac180e574139" exitCode=0 Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.767923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" event={"ID":"bb4ea3d2-63c7-4829-b6b8-8112476493fa","Type":"ContainerDied","Data":"f5715a5c2dbadf161e906011876459ee94ba9a1523d9455c1d55ac180e574139"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.767951 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" event={"ID":"bb4ea3d2-63c7-4829-b6b8-8112476493fa","Type":"ContainerStarted","Data":"5e519680bc7c55a0ecc86ba490dcbb3969a313e5b2bc06cbb94da1a485a26d3b"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.770903 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerStarted","Data":"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.770942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerStarted","Data":"eaaf72feb767aa6ebf839a27a99ea70de0682afc00ba0ae8e8e10dabfcaf1864"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.772472 4824 generic.go:334] "Generic (PLEG): container finished" podID="08b8c363-a8c1-4dfe-8fd1-1af2d055a305" containerID="5f4a1e6bcd0c138e59cd1e69d47cce9046bae28ab6b9ce57bd7be2ad0c6a4e61" exitCode=0 Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.772517 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6vgh6" event={"ID":"08b8c363-a8c1-4dfe-8fd1-1af2d055a305","Type":"ContainerDied","Data":"5f4a1e6bcd0c138e59cd1e69d47cce9046bae28ab6b9ce57bd7be2ad0c6a4e61"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.773759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerStarted","Data":"953a68ff3f5b22fd88f3a1a2babc253f6f9d0648a9b5bca14c07b85059d64e1a"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.773782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerStarted","Data":"5dc758f1e11243d2a910301ab89754d05c98dfa5446bb49c248a5833a44c6ce9"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.774846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5l2gg" event={"ID":"a7e75011-d633-4ccc-951c-d019cb2100f9","Type":"ContainerStarted","Data":"2534da156868c09c8dd8dee0424c34fab5c7b1401695a79ef623bde9f1ef00f0"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.774870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5l2gg" event={"ID":"a7e75011-d633-4ccc-951c-d019cb2100f9","Type":"ContainerStarted","Data":"f73eda2acba113b106a0d01407072387e56c3b86b9c9c6135b58a1185903e93b"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.778015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerStarted","Data":"ff804ebe93733fa9e697132922f8e30c6ee75d058b7cb8afecec7da791271c35"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.778041 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerStarted","Data":"2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.802124 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5l2gg" podStartSLOduration=7.8021119169999995 podStartE2EDuration="7.802111917s" podCreationTimestamp="2026-01-21 11:26:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:42.795007916 +0000 UTC m=+945.088037209" watchObservedRunningTime="2026-01-21 11:26:42.802111917 +0000 UTC m=+945.095141209" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.486787 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-659565f76f-jdg2q"] Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.489747 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.491454 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.492619 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.521347 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-659565f76f-jdg2q"] Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-internal-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590515 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-httpd-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590545 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-ovndb-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6dqp\" (UniqueName: \"kubernetes.io/projected/b6f2ba03-746e-465c-bde2-c917129fb125-kube-api-access-c6dqp\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-public-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590674 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-combined-ca-bundle\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-internal-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692292 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-httpd-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-ovndb-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692408 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6dqp\" (UniqueName: \"kubernetes.io/projected/b6f2ba03-746e-465c-bde2-c917129fb125-kube-api-access-c6dqp\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-public-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692443 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-combined-ca-bundle\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.699753 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-combined-ca-bundle\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.699877 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-ovndb-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.700008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-internal-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.701064 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-public-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.701601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.708427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-httpd-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.711473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6dqp\" (UniqueName: \"kubernetes.io/projected/b6f2ba03-746e-465c-bde2-c917129fb125-kube-api-access-c6dqp\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.784317 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerStarted","Data":"eb6d7862490b4afbdbbe422c8984287dc94d6e4a39261b3846440fb920c542fb"} Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.784399 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.785777 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" event={"ID":"bb4ea3d2-63c7-4829-b6b8-8112476493fa","Type":"ContainerStarted","Data":"25b08d9ba24d1cbae3803365bca42ff0e044f0f812800389a684ef8eb1362dbe"} Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.786282 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.787419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerStarted","Data":"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686"} Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.789307 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerStarted","Data":"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128"} Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.790797 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerStarted","Data":"77d449bfcfde658bccf9c84cc9cff76978535cecd3aaeeb5e306a982392e5cf4"} Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.801537 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-79bfcdcfbd-s2v5h" podStartSLOduration=2.801521821 podStartE2EDuration="2.801521821s" podCreationTimestamp="2026-01-21 11:26:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:43.797167158 +0000 UTC m=+946.090196451" watchObservedRunningTime="2026-01-21 11:26:43.801521821 +0000 UTC m=+946.094551113" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.824540 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.824524189 podStartE2EDuration="3.824524189s" podCreationTimestamp="2026-01-21 11:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:43.819076985 +0000 UTC m=+946.112106277" watchObservedRunningTime="2026-01-21 11:26:43.824524189 +0000 UTC m=+946.117553481" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.836637 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.855318 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" podStartSLOduration=2.855302655 podStartE2EDuration="2.855302655s" podCreationTimestamp="2026-01-21 11:26:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:43.853986693 +0000 UTC m=+946.147015985" watchObservedRunningTime="2026-01-21 11:26:43.855302655 +0000 UTC m=+946.148331947" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.858329 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.858318923 podStartE2EDuration="3.858318923s" podCreationTimestamp="2026-01-21 11:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:43.841981959 +0000 UTC m=+946.135011251" watchObservedRunningTime="2026-01-21 11:26:43.858318923 +0000 UTC m=+946.151348226" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.148327 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.208497 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts\") pod \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.208639 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data\") pod \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.208728 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs\") pod \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.208774 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkb9f\" (UniqueName: \"kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f\") pod \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.208803 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle\") pod \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.209989 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs" (OuterVolumeSpecName: "logs") pod "08b8c363-a8c1-4dfe-8fd1-1af2d055a305" (UID: "08b8c363-a8c1-4dfe-8fd1-1af2d055a305"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.213224 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts" (OuterVolumeSpecName: "scripts") pod "08b8c363-a8c1-4dfe-8fd1-1af2d055a305" (UID: "08b8c363-a8c1-4dfe-8fd1-1af2d055a305"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.214142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f" (OuterVolumeSpecName: "kube-api-access-gkb9f") pod "08b8c363-a8c1-4dfe-8fd1-1af2d055a305" (UID: "08b8c363-a8c1-4dfe-8fd1-1af2d055a305"). InnerVolumeSpecName "kube-api-access-gkb9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.273616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data" (OuterVolumeSpecName: "config-data") pod "08b8c363-a8c1-4dfe-8fd1-1af2d055a305" (UID: "08b8c363-a8c1-4dfe-8fd1-1af2d055a305"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.290128 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08b8c363-a8c1-4dfe-8fd1-1af2d055a305" (UID: "08b8c363-a8c1-4dfe-8fd1-1af2d055a305"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.312970 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.313064 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkb9f\" (UniqueName: \"kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.313119 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.313164 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.313208 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.532711 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-659565f76f-jdg2q"] Jan 21 11:26:44 crc kubenswrapper[4824]: W0121 11:26:44.535620 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6f2ba03_746e_465c_bde2_c917129fb125.slice/crio-e0b06e6954ac89ea6fb74405de5e37daee67b3757a4ece6c64e7e321e47249be WatchSource:0}: Error finding container e0b06e6954ac89ea6fb74405de5e37daee67b3757a4ece6c64e7e321e47249be: Status 404 returned error can't find the container with id e0b06e6954ac89ea6fb74405de5e37daee67b3757a4ece6c64e7e321e47249be Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.799623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6vgh6" event={"ID":"08b8c363-a8c1-4dfe-8fd1-1af2d055a305","Type":"ContainerDied","Data":"97721ccfe8abdf2324794ac67debd67a8fc655d3d7ab538bf2cb4b35e35f8386"} Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.799672 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97721ccfe8abdf2324794ac67debd67a8fc655d3d7ab538bf2cb4b35e35f8386" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.799638 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.801196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-659565f76f-jdg2q" event={"ID":"b6f2ba03-746e-465c-bde2-c917129fb125","Type":"ContainerStarted","Data":"966a3b02cb0a14177696e58a750fbeec6fd6ca5d319be90ce60f05fc0315f36f"} Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.801233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-659565f76f-jdg2q" event={"ID":"b6f2ba03-746e-465c-bde2-c917129fb125","Type":"ContainerStarted","Data":"e0b06e6954ac89ea6fb74405de5e37daee67b3757a4ece6c64e7e321e47249be"} Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.803552 4824 generic.go:334] "Generic (PLEG): container finished" podID="cadf2d38-24f9-43ab-8e29-803d97a0ff06" containerID="5dcf0246f4ab2f2f4f3230d6a47ec6ac1102fcb34b4edb4d89a73c99d792f5ff" exitCode=0 Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.803602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jrsn8" event={"ID":"cadf2d38-24f9-43ab-8e29-803d97a0ff06","Type":"ContainerDied","Data":"5dcf0246f4ab2f2f4f3230d6a47ec6ac1102fcb34b4edb4d89a73c99d792f5ff"} Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.805507 4824 generic.go:334] "Generic (PLEG): container finished" podID="a7e75011-d633-4ccc-951c-d019cb2100f9" containerID="2534da156868c09c8dd8dee0424c34fab5c7b1401695a79ef623bde9f1ef00f0" exitCode=0 Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.805620 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5l2gg" event={"ID":"a7e75011-d633-4ccc-951c-d019cb2100f9","Type":"ContainerDied","Data":"2534da156868c09c8dd8dee0424c34fab5c7b1401695a79ef623bde9f1ef00f0"} Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.924709 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8654bbf998-p8crm"] Jan 21 11:26:44 crc kubenswrapper[4824]: E0121 11:26:44.925076 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b8c363-a8c1-4dfe-8fd1-1af2d055a305" containerName="placement-db-sync" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.925095 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b8c363-a8c1-4dfe-8fd1-1af2d055a305" containerName="placement-db-sync" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.925286 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="08b8c363-a8c1-4dfe-8fd1-1af2d055a305" containerName="placement-db-sync" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.926251 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.929739 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.929986 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.930130 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.930203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n8hrk" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.930232 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.933406 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8654bbf998-p8crm"] Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.027892 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-combined-ca-bundle\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.027993 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-scripts\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.028022 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-public-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.028058 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-config-data\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.028101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23477ba-5a41-44da-a8b7-6606595b9726-logs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.028120 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlbpz\" (UniqueName: \"kubernetes.io/projected/c23477ba-5a41-44da-a8b7-6606595b9726-kube-api-access-dlbpz\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.028140 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-internal-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-scripts\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129569 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-public-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-config-data\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23477ba-5a41-44da-a8b7-6606595b9726-logs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlbpz\" (UniqueName: \"kubernetes.io/projected/c23477ba-5a41-44da-a8b7-6606595b9726-kube-api-access-dlbpz\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129730 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-internal-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-combined-ca-bundle\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.130587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23477ba-5a41-44da-a8b7-6606595b9726-logs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.133684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-config-data\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.136202 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-public-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.139329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-scripts\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.142389 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-internal-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.144400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlbpz\" (UniqueName: \"kubernetes.io/projected/c23477ba-5a41-44da-a8b7-6606595b9726-kube-api-access-dlbpz\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.157369 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-combined-ca-bundle\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.250627 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.719250 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.814872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-659565f76f-jdg2q" event={"ID":"b6f2ba03-746e-465c-bde2-c917129fb125","Type":"ContainerStarted","Data":"4b3e1a4d84baaeed6c9761ff58fa7b697d120977b82c7b9024880b67d908aeed"} Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.815790 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.833828 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-659565f76f-jdg2q" podStartSLOduration=2.833813488 podStartE2EDuration="2.833813488s" podCreationTimestamp="2026-01-21 11:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:45.827439735 +0000 UTC m=+948.120469027" watchObservedRunningTime="2026-01-21 11:26:45.833813488 +0000 UTC m=+948.126842780" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.065653 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.065698 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.406551 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553440 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lj6r\" (UniqueName: \"kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553483 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553573 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553605 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553668 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553688 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.561259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.561425 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.566203 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts" (OuterVolumeSpecName: "scripts") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.568491 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r" (OuterVolumeSpecName: "kube-api-access-8lj6r") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "kube-api-access-8lj6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.576268 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data" (OuterVolumeSpecName: "config-data") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.594180 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655225 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655253 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lj6r\" (UniqueName: \"kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655265 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655273 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655280 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655287 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.822265 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5l2gg" event={"ID":"a7e75011-d633-4ccc-951c-d019cb2100f9","Type":"ContainerDied","Data":"f73eda2acba113b106a0d01407072387e56c3b86b9c9c6135b58a1185903e93b"} Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.822311 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f73eda2acba113b106a0d01407072387e56c3b86b9c9c6135b58a1185903e93b" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.822429 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.984773 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c5fc769c9-spmzr"] Jan 21 11:26:46 crc kubenswrapper[4824]: E0121 11:26:46.985408 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e75011-d633-4ccc-951c-d019cb2100f9" containerName="keystone-bootstrap" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.985422 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e75011-d633-4ccc-951c-d019cb2100f9" containerName="keystone-bootstrap" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.985585 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e75011-d633-4ccc-951c-d019cb2100f9" containerName="keystone-bootstrap" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.986065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.989312 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.989476 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxhzv" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.989604 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.989718 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.989818 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.990271 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.998563 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c5fc769c9-spmzr"] Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-public-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-internal-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-fernet-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-scripts\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165594 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-combined-ca-bundle\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165653 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-credential-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v487x\" (UniqueName: \"kubernetes.io/projected/7d943b5e-d05f-445a-888d-7ebe340468b9-kube-api-access-v487x\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165719 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-config-data\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267414 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-combined-ca-bundle\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-credential-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267538 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v487x\" (UniqueName: \"kubernetes.io/projected/7d943b5e-d05f-445a-888d-7ebe340468b9-kube-api-access-v487x\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267563 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-config-data\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-public-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-internal-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-fernet-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-scripts\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.272072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-credential-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.273229 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-public-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.273669 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-scripts\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.273752 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-config-data\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.274029 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-internal-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.274171 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-combined-ca-bundle\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.279255 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-fernet-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.281428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v487x\" (UniqueName: \"kubernetes.io/projected/7d943b5e-d05f-445a-888d-7ebe340468b9-kube-api-access-v487x\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.304929 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.844442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jrsn8" event={"ID":"cadf2d38-24f9-43ab-8e29-803d97a0ff06","Type":"ContainerDied","Data":"f8ce50a0aaa511cfcd5c91e987379f42c2650248768a77a65a4355ed457a06c8"} Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.844627 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8ce50a0aaa511cfcd5c91e987379f42c2650248768a77a65a4355ed457a06c8" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.879314 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.980836 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrdwp\" (UniqueName: \"kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp\") pod \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.981030 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data\") pod \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.981075 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle\") pod \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.984410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp" (OuterVolumeSpecName: "kube-api-access-xrdwp") pod "cadf2d38-24f9-43ab-8e29-803d97a0ff06" (UID: "cadf2d38-24f9-43ab-8e29-803d97a0ff06"). InnerVolumeSpecName "kube-api-access-xrdwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.001110 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cadf2d38-24f9-43ab-8e29-803d97a0ff06" (UID: "cadf2d38-24f9-43ab-8e29-803d97a0ff06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.027610 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data" (OuterVolumeSpecName: "config-data") pod "cadf2d38-24f9-43ab-8e29-803d97a0ff06" (UID: "cadf2d38-24f9-43ab-8e29-803d97a0ff06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.082522 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.082543 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.082552 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrdwp\" (UniqueName: \"kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.150075 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8654bbf998-p8crm"] Jan 21 11:26:48 crc kubenswrapper[4824]: W0121 11:26:48.158563 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d943b5e_d05f_445a_888d_7ebe340468b9.slice/crio-779d43c0e7120424033bc4414b128725901d8d0023671ffa45ed85a0227c9c69 WatchSource:0}: Error finding container 779d43c0e7120424033bc4414b128725901d8d0023671ffa45ed85a0227c9c69: Status 404 returned error can't find the container with id 779d43c0e7120424033bc4414b128725901d8d0023671ffa45ed85a0227c9c69 Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.159922 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c5fc769c9-spmzr"] Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.174450 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:26:48 crc kubenswrapper[4824]: E0121 11:26:48.174757 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cadf2d38-24f9-43ab-8e29-803d97a0ff06" containerName="heat-db-sync" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.174775 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cadf2d38-24f9-43ab-8e29-803d97a0ff06" containerName="heat-db-sync" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.174965 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cadf2d38-24f9-43ab-8e29-803d97a0ff06" containerName="heat-db-sync" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.176674 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.186275 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.285635 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.285900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.286007 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9j78\" (UniqueName: \"kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.387435 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.387524 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.387609 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9j78\" (UniqueName: \"kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.388218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.388497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.404589 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9j78\" (UniqueName: \"kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.502022 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.851239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerStarted","Data":"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.852190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c5fc769c9-spmzr" event={"ID":"7d943b5e-d05f-445a-888d-7ebe340468b9","Type":"ContainerStarted","Data":"97fb464b408f5ceff487705be319b8cd670592be37d52a0443879365b61921a9"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.852218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c5fc769c9-spmzr" event={"ID":"7d943b5e-d05f-445a-888d-7ebe340468b9","Type":"ContainerStarted","Data":"779d43c0e7120424033bc4414b128725901d8d0023671ffa45ed85a0227c9c69"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.852980 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854367 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8654bbf998-p8crm" event={"ID":"c23477ba-5a41-44da-a8b7-6606595b9726","Type":"ContainerStarted","Data":"a33d5c3f52560f39a45b643e1e7362cec3d1ffa8698489a0cd9d0e49beebc4f3"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8654bbf998-p8crm" event={"ID":"c23477ba-5a41-44da-a8b7-6606595b9726","Type":"ContainerStarted","Data":"db3de21ed146f19802de452a644645642de1b010142ab29da18a4e42e355a6fe"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854555 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854579 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8654bbf998-p8crm" event={"ID":"c23477ba-5a41-44da-a8b7-6606595b9726","Type":"ContainerStarted","Data":"d922ef122df4ca799eb3f9e4b5e4b6c48e353cdfcff00495934a761838cd1cd6"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854592 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.868382 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c5fc769c9-spmzr" podStartSLOduration=2.868369402 podStartE2EDuration="2.868369402s" podCreationTimestamp="2026-01-21 11:26:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:48.864564345 +0000 UTC m=+951.157593637" watchObservedRunningTime="2026-01-21 11:26:48.868369402 +0000 UTC m=+951.161398695" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.883380 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-8654bbf998-p8crm" podStartSLOduration=4.883369656 podStartE2EDuration="4.883369656s" podCreationTimestamp="2026-01-21 11:26:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:48.879331099 +0000 UTC m=+951.172360391" watchObservedRunningTime="2026-01-21 11:26:48.883369656 +0000 UTC m=+951.176398949" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.966346 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:26:48 crc kubenswrapper[4824]: W0121 11:26:48.969857 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d8c8cff_4340_4eb4_94b7_062968a14848.slice/crio-2dec804f6fee46d1464f08de4ca7da8580e7f171b751ac42a7b92ec863162d9b WatchSource:0}: Error finding container 2dec804f6fee46d1464f08de4ca7da8580e7f171b751ac42a7b92ec863162d9b: Status 404 returned error can't find the container with id 2dec804f6fee46d1464f08de4ca7da8580e7f171b751ac42a7b92ec863162d9b Jan 21 11:26:49 crc kubenswrapper[4824]: I0121 11:26:49.875471 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerID="d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a" exitCode=0 Jan 21 11:26:49 crc kubenswrapper[4824]: I0121 11:26:49.876695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerDied","Data":"d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a"} Jan 21 11:26:49 crc kubenswrapper[4824]: I0121 11:26:49.876734 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerStarted","Data":"2dec804f6fee46d1464f08de4ca7da8580e7f171b751ac42a7b92ec863162d9b"} Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.051790 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.052040 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.063832 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.063870 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.073280 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.079262 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.092407 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.094792 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.513101 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.565592 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.565823 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="dnsmasq-dns" containerID="cri-o://56442b913a4db71125dc75408dc50b059e8519640a2cc312a0c8277e0b1aefab" gracePeriod=10 Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893009 4824 generic.go:334] "Generic (PLEG): container finished" podID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerID="56442b913a4db71125dc75408dc50b059e8519640a2cc312a0c8277e0b1aefab" exitCode=0 Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" event={"ID":"8214df67-a984-4a5b-bcbb-6b16ab42f5fe","Type":"ContainerDied","Data":"56442b913a4db71125dc75408dc50b059e8519640a2cc312a0c8277e0b1aefab"} Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893813 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893838 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893848 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893856 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 21 11:26:52 crc kubenswrapper[4824]: I0121 11:26:52.012344 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.140:5353: connect: connection refused" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.430854 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.538158 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596323 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596457 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596503 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8x9b\" (UniqueName: \"kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596587 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596619 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.606581 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.614630 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b" (OuterVolumeSpecName: "kube-api-access-x8x9b") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "kube-api-access-x8x9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.634247 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config" (OuterVolumeSpecName: "config") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.634284 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.638617 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.642602 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.650005 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.666256 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698469 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698496 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8x9b\" (UniqueName: \"kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698508 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698516 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698525 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698532 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.799551 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.909266 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.909591 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" event={"ID":"8214df67-a984-4a5b-bcbb-6b16ab42f5fe","Type":"ContainerDied","Data":"a7ed7bee402dc3db5565fe550b060e37c1771f55a94528a7a39dbd2741c1ec27"} Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.909632 4824 scope.go:117] "RemoveContainer" containerID="56442b913a4db71125dc75408dc50b059e8519640a2cc312a0c8277e0b1aefab" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.954348 4824 scope.go:117] "RemoveContainer" containerID="7af315bc544e0ed568b3db042720a7c28e2f59b1fb260d2208c816705728d7f5" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.978680 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.984401 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.062074 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" path="/var/lib/kubelet/pods/8214df67-a984-4a5b-bcbb-6b16ab42f5fe/volumes" Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926375 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerStarted","Data":"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b"} Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926508 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-central-agent" containerID="cri-o://c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c" gracePeriod=30 Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926621 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-notification-agent" containerID="cri-o://af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128" gracePeriod=30 Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926666 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="proxy-httpd" containerID="cri-o://c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b" gracePeriod=30 Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926689 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926769 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="sg-core" containerID="cri-o://be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee" gracePeriod=30 Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.930845 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerID="d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943" exitCode=0 Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.931949 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerDied","Data":"d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943"} Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.948226 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.473406197 podStartE2EDuration="33.948216281s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="2026-01-21 11:26:22.402277514 +0000 UTC m=+924.695306806" lastFinishedPulling="2026-01-21 11:26:53.877087598 +0000 UTC m=+956.170116890" observedRunningTime="2026-01-21 11:26:54.945212356 +0000 UTC m=+957.238241648" watchObservedRunningTime="2026-01-21 11:26:54.948216281 +0000 UTC m=+957.241245573" Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.938596 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rlp2t" event={"ID":"0a6c0628-4522-4bb7-8a82-cc2e019eca2a","Type":"ContainerStarted","Data":"b5f108624f476c15b99bcdcb7c95850ef2819568712069d3c2536bdcb6daa49b"} Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.952367 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-rlp2t" podStartSLOduration=2.6327185650000002 podStartE2EDuration="33.952354966s" podCreationTimestamp="2026-01-21 11:26:22 +0000 UTC" firstStartedPulling="2026-01-21 11:26:23.218138454 +0000 UTC m=+925.511167746" lastFinishedPulling="2026-01-21 11:26:54.537774854 +0000 UTC m=+956.830804147" observedRunningTime="2026-01-21 11:26:55.951670775 +0000 UTC m=+958.244700067" watchObservedRunningTime="2026-01-21 11:26:55.952354966 +0000 UTC m=+958.245384257" Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.961988 4824 generic.go:334] "Generic (PLEG): container finished" podID="bbefda93-dc56-43c5-b131-56af96ca832b" containerID="c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b" exitCode=0 Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.962019 4824 generic.go:334] "Generic (PLEG): container finished" podID="bbefda93-dc56-43c5-b131-56af96ca832b" containerID="be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee" exitCode=2 Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.962028 4824 generic.go:334] "Generic (PLEG): container finished" podID="bbefda93-dc56-43c5-b131-56af96ca832b" containerID="c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c" exitCode=0 Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.962055 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerDied","Data":"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b"} Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.962104 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerDied","Data":"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee"} Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.962115 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerDied","Data":"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c"} Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.964446 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerStarted","Data":"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4"} Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.982306 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dn2np" podStartSLOduration=2.405619777 podStartE2EDuration="7.982293647s" podCreationTimestamp="2026-01-21 11:26:48 +0000 UTC" firstStartedPulling="2026-01-21 11:26:49.877638805 +0000 UTC m=+952.170668097" lastFinishedPulling="2026-01-21 11:26:55.454312674 +0000 UTC m=+957.747341967" observedRunningTime="2026-01-21 11:26:55.980171524 +0000 UTC m=+958.273200817" watchObservedRunningTime="2026-01-21 11:26:55.982293647 +0000 UTC m=+958.275322940" Jan 21 11:26:56 crc kubenswrapper[4824]: I0121 11:26:56.972762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ngq69" event={"ID":"262733b9-4179-49e3-aee9-b62197cc89ba","Type":"ContainerStarted","Data":"4c9c6f12e02388718e6b9fff3020bdecb9cf28f3ef2f988918e2aeb78dac7d68"} Jan 21 11:26:56 crc kubenswrapper[4824]: I0121 11:26:56.986255 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ngq69" podStartSLOduration=1.8551470490000002 podStartE2EDuration="34.986243206s" podCreationTimestamp="2026-01-21 11:26:22 +0000 UTC" firstStartedPulling="2026-01-21 11:26:23.395650882 +0000 UTC m=+925.688680175" lastFinishedPulling="2026-01-21 11:26:56.52674704 +0000 UTC m=+958.819776332" observedRunningTime="2026-01-21 11:26:56.983586665 +0000 UTC m=+959.276615946" watchObservedRunningTime="2026-01-21 11:26:56.986243206 +0000 UTC m=+959.279272498" Jan 21 11:26:57 crc kubenswrapper[4824]: I0121 11:26:57.979659 4824 generic.go:334] "Generic (PLEG): container finished" podID="262733b9-4179-49e3-aee9-b62197cc89ba" containerID="4c9c6f12e02388718e6b9fff3020bdecb9cf28f3ef2f988918e2aeb78dac7d68" exitCode=0 Jan 21 11:26:57 crc kubenswrapper[4824]: I0121 11:26:57.979753 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ngq69" event={"ID":"262733b9-4179-49e3-aee9-b62197cc89ba","Type":"ContainerDied","Data":"4c9c6f12e02388718e6b9fff3020bdecb9cf28f3ef2f988918e2aeb78dac7d68"} Jan 21 11:26:57 crc kubenswrapper[4824]: I0121 11:26:57.980973 4824 generic.go:334] "Generic (PLEG): container finished" podID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" containerID="b5f108624f476c15b99bcdcb7c95850ef2819568712069d3c2536bdcb6daa49b" exitCode=0 Jan 21 11:26:57 crc kubenswrapper[4824]: I0121 11:26:57.981012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rlp2t" event={"ID":"0a6c0628-4522-4bb7-8a82-cc2e019eca2a","Type":"ContainerDied","Data":"b5f108624f476c15b99bcdcb7c95850ef2819568712069d3c2536bdcb6daa49b"} Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.308506 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.489621 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490008 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490040 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgmtk\" (UniqueName: \"kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490099 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490171 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490262 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490288 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490493 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490902 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490897 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.494304 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts" (OuterVolumeSpecName: "scripts") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.494395 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk" (OuterVolumeSpecName: "kube-api-access-sgmtk") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "kube-api-access-sgmtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.502505 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.502560 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.509616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.533679 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.536144 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.552572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data" (OuterVolumeSpecName: "config-data") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592867 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgmtk\" (UniqueName: \"kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592902 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592912 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592920 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592928 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592935 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.989804 4824 generic.go:334] "Generic (PLEG): container finished" podID="bbefda93-dc56-43c5-b131-56af96ca832b" containerID="af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128" exitCode=0 Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.989873 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.989917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerDied","Data":"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128"} Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.989976 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerDied","Data":"9fa11b86f012b31aa61fb0528e706362f80104db47ab25f797420eb04f92a3fa"} Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.990001 4824 scope.go:117] "RemoveContainer" containerID="c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.008047 4824 scope.go:117] "RemoveContainer" containerID="be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.021348 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.024647 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034266 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034578 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="sg-core" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034597 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="sg-core" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034618 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="dnsmasq-dns" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034624 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="dnsmasq-dns" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034637 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-central-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034644 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-central-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034656 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="init" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034661 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="init" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034674 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-notification-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034679 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-notification-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034690 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="proxy-httpd" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034695 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="proxy-httpd" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034850 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-central-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034874 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-notification-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034893 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="sg-core" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034905 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="proxy-httpd" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034919 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="dnsmasq-dns" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.036325 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.038799 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.038975 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.046343 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.047575 4824 scope.go:117] "RemoveContainer" containerID="af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.088695 4824 scope.go:117] "RemoveContainer" containerID="c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.103581 4824 scope.go:117] "RemoveContainer" containerID="c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.104245 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b\": container with ID starting with c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b not found: ID does not exist" containerID="c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.104283 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b"} err="failed to get container status \"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b\": rpc error: code = NotFound desc = could not find container \"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b\": container with ID starting with c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b not found: ID does not exist" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.104328 4824 scope.go:117] "RemoveContainer" containerID="be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.104638 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee\": container with ID starting with be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee not found: ID does not exist" containerID="be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.104667 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee"} err="failed to get container status \"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee\": rpc error: code = NotFound desc = could not find container \"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee\": container with ID starting with be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee not found: ID does not exist" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.104686 4824 scope.go:117] "RemoveContainer" containerID="af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.104980 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128\": container with ID starting with af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128 not found: ID does not exist" containerID="af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.105014 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128"} err="failed to get container status \"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128\": rpc error: code = NotFound desc = could not find container \"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128\": container with ID starting with af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128 not found: ID does not exist" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.105038 4824 scope.go:117] "RemoveContainer" containerID="c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.105308 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c\": container with ID starting with c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c not found: ID does not exist" containerID="c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.105351 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c"} err="failed to get container status \"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c\": rpc error: code = NotFound desc = could not find container \"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c\": container with ID starting with c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c not found: ID does not exist" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200361 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200609 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmjb9\" (UniqueName: \"kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200652 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200690 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200791 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200811 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302258 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmjb9\" (UniqueName: \"kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302416 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302436 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302456 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302825 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.303040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.308038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.308434 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.314134 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.315351 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.317879 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmjb9\" (UniqueName: \"kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.375402 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.381163 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.385433 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505518 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505739 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505801 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505832 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505864 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data\") pod \"262733b9-4179-49e3-aee9-b62197cc89ba\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505887 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle\") pod \"262733b9-4179-49e3-aee9-b62197cc89ba\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505916 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505972 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flpqh\" (UniqueName: \"kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh\") pod \"262733b9-4179-49e3-aee9-b62197cc89ba\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.506033 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msmnx\" (UniqueName: \"kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.506952 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.509791 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.510198 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts" (OuterVolumeSpecName: "scripts") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.510359 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx" (OuterVolumeSpecName: "kube-api-access-msmnx") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "kube-api-access-msmnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.510780 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh" (OuterVolumeSpecName: "kube-api-access-flpqh") pod "262733b9-4179-49e3-aee9-b62197cc89ba" (UID: "262733b9-4179-49e3-aee9-b62197cc89ba"). InnerVolumeSpecName "kube-api-access-flpqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.510846 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "262733b9-4179-49e3-aee9-b62197cc89ba" (UID: "262733b9-4179-49e3-aee9-b62197cc89ba"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.524943 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.534166 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "262733b9-4179-49e3-aee9-b62197cc89ba" (UID: "262733b9-4179-49e3-aee9-b62197cc89ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.542048 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data" (OuterVolumeSpecName: "config-data") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607883 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607910 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607922 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607931 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607939 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607947 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607967 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607975 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flpqh\" (UniqueName: \"kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607984 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msmnx\" (UniqueName: \"kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.749396 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:59 crc kubenswrapper[4824]: W0121 11:26:59.750870 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbae23da5_948a_4883_b88d_d6ed5d6ccba7.slice/crio-b8cc173acd912ad690d3e7aa92c24c20b537f220a873ef96f5148fc3b668366a WatchSource:0}: Error finding container b8cc173acd912ad690d3e7aa92c24c20b537f220a873ef96f5148fc3b668366a: Status 404 returned error can't find the container with id b8cc173acd912ad690d3e7aa92c24c20b537f220a873ef96f5148fc3b668366a Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.997088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rlp2t" event={"ID":"0a6c0628-4522-4bb7-8a82-cc2e019eca2a","Type":"ContainerDied","Data":"877238fc09eb99e400629467da99c2ba4a6e53c0afc16677b1308d3c52849ae9"} Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.997118 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.997137 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="877238fc09eb99e400629467da99c2ba4a6e53c0afc16677b1308d3c52849ae9" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.999826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ngq69" event={"ID":"262733b9-4179-49e3-aee9-b62197cc89ba","Type":"ContainerDied","Data":"2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf"} Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.999850 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.999888 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ngq69" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.002319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerStarted","Data":"b8cc173acd912ad690d3e7aa92c24c20b537f220a873ef96f5148fc3b668366a"} Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.042352 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.058644 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" path="/var/lib/kubelet/pods/bbefda93-dc56-43c5-b131-56af96ca832b/volumes" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.082019 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.236144 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:00 crc kubenswrapper[4824]: E0121 11:27:00.236464 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" containerName="barbican-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.236478 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" containerName="barbican-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: E0121 11:27:00.236486 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" containerName="cinder-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.236492 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" containerName="cinder-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.236666 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" containerName="barbican-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.236699 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" containerName="cinder-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.238431 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.242892 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2xjd5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.242985 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.243092 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.250524 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.261103 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.295670 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.325133 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.368639 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428542 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgb7x\" (UniqueName: \"kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428593 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428703 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428792 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428874 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428899 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428940 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428988 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.429012 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.429038 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmmwh\" (UniqueName: \"kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.429082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.441295 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.442546 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.445381 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.452641 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.533684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.533971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534028 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534075 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534100 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534120 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534141 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmmwh\" (UniqueName: \"kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534173 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534239 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534266 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgb7x\" (UniqueName: \"kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534305 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534495 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534744 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.535360 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.535947 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.536462 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.537198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.543461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.543647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.546563 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.554581 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.568406 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmmwh\" (UniqueName: \"kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.568469 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7cf777545f-f8drb"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.569790 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.570439 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.571487 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7cf777545f-f8drb"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.576613 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.576786 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8q76f" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.576916 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.594274 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgb7x\" (UniqueName: \"kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.596032 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.599411 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.606906 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.613257 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635492 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2qhr\" (UniqueName: \"kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635681 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.663496 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.664182 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.676472 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.678859 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.692528 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-combined-ca-bundle\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737806 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737906 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b84edbe-4485-4027-bfc4-cebbb1b10543-logs\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738014 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdbsk\" (UniqueName: \"kubernetes.io/projected/283e6982-7444-48c7-b7ac-ed4210329a08-kube-api-access-wdbsk\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data-custom\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2qhr\" (UniqueName: \"kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data-custom\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738166 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-combined-ca-bundle\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738180 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/283e6982-7444-48c7-b7ac-ed4210329a08-logs\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738205 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738232 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spbcb\" (UniqueName: \"kubernetes.io/projected/8b84edbe-4485-4027-bfc4-cebbb1b10543-kube-api-access-spbcb\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.739496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.746800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.748058 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.749500 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.749685 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.758329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.763796 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.765168 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.766135 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2qhr\" (UniqueName: \"kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.770883 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.774791 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.778302 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.840855 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.840941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data-custom\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.840983 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841012 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data-custom\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-combined-ca-bundle\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/283e6982-7444-48c7-b7ac-ed4210329a08-logs\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spbcb\" (UniqueName: \"kubernetes.io/projected/8b84edbe-4485-4027-bfc4-cebbb1b10543-kube-api-access-spbcb\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841142 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841157 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-combined-ca-bundle\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841206 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841233 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4ttl\" (UniqueName: \"kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b84edbe-4485-4027-bfc4-cebbb1b10543-logs\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdbsk\" (UniqueName: \"kubernetes.io/projected/283e6982-7444-48c7-b7ac-ed4210329a08-kube-api-access-wdbsk\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/283e6982-7444-48c7-b7ac-ed4210329a08-logs\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.842078 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b84edbe-4485-4027-bfc4-cebbb1b10543-logs\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.846461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.847499 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data-custom\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.848617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-combined-ca-bundle\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.848793 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-combined-ca-bundle\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.849419 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data-custom\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.849878 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.857140 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spbcb\" (UniqueName: \"kubernetes.io/projected/8b84edbe-4485-4027-bfc4-cebbb1b10543-kube-api-access-spbcb\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.857940 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdbsk\" (UniqueName: \"kubernetes.io/projected/283e6982-7444-48c7-b7ac-ed4210329a08-kube-api-access-wdbsk\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.898499 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.926707 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942080 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942127 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942163 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942187 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942894 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942945 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943551 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943001 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnlw6\" (UniqueName: \"kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943800 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943824 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943844 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4ttl\" (UniqueName: \"kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.944363 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.960852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4ttl\" (UniqueName: \"kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.006764 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.013937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerStarted","Data":"7b2e891014d0053a1d16ec02f1fc280a3a0cbb96983be7316f8030e04e57c2be"} Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.044946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.045018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.045122 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnlw6\" (UniqueName: \"kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.045184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.045211 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.046182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.049606 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.053151 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.054160 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.062419 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnlw6\" (UniqueName: \"kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.102462 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.118071 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.154689 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.234406 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:01 crc kubenswrapper[4824]: W0121 11:27:01.302027 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5721831_a6aa_4d42_b62a_041dbe5c23ad.slice/crio-ccfadddd012e5287b137e7bb05a0f3a68e04a10bc60fc278e3e35eab10438b65 WatchSource:0}: Error finding container ccfadddd012e5287b137e7bb05a0f3a68e04a10bc60fc278e3e35eab10438b65: Status 404 returned error can't find the container with id ccfadddd012e5287b137e7bb05a0f3a68e04a10bc60fc278e3e35eab10438b65 Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.338859 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7cf777545f-f8drb"] Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.438535 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5"] Jan 21 11:27:01 crc kubenswrapper[4824]: W0121 11:27:01.440899 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b84edbe_4485_4027_bfc4_cebbb1b10543.slice/crio-55fc42714f1e8c2485a2f2ab2991b189f14890bb62beec8d56c4a223261fc69c WatchSource:0}: Error finding container 55fc42714f1e8c2485a2f2ab2991b189f14890bb62beec8d56c4a223261fc69c: Status 404 returned error can't find the container with id 55fc42714f1e8c2485a2f2ab2991b189f14890bb62beec8d56c4a223261fc69c Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.490768 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:01 crc kubenswrapper[4824]: W0121 11:27:01.502357 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod661c81f6_6854_416e_aea7_1344a635f2db.slice/crio-e5f93c719de6d721a85fa8dea128d5a96c33c0da5f5e176c82969aadb4af41ac WatchSource:0}: Error finding container e5f93c719de6d721a85fa8dea128d5a96c33c0da5f5e176c82969aadb4af41ac: Status 404 returned error can't find the container with id e5f93c719de6d721a85fa8dea128d5a96c33c0da5f5e176c82969aadb4af41ac Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.715221 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:01 crc kubenswrapper[4824]: W0121 11:27:01.777363 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ed8e1b1_57c3_4b01_bbd7_9dc5125f4f94.slice/crio-62531d41bed5d935b440d701f3d3a82064300927572a77364b43daf9ada74a10 WatchSource:0}: Error finding container 62531d41bed5d935b440d701f3d3a82064300927572a77364b43daf9ada74a10: Status 404 returned error can't find the container with id 62531d41bed5d935b440d701f3d3a82064300927572a77364b43daf9ada74a10 Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.024329 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerStarted","Data":"41e3cbe62a3cfe1f1349c3e05b9d5169c6ca9dd78e1f915854df40b91a504d55"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.024574 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerStarted","Data":"62531d41bed5d935b440d701f3d3a82064300927572a77364b43daf9ada74a10"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.026163 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerStarted","Data":"3b7c3d7bba0374d53d0ac03b9cb2de415ba41cc2525693fb788df9d2d6551fe4"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.028256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerStarted","Data":"ccfadddd012e5287b137e7bb05a0f3a68e04a10bc60fc278e3e35eab10438b65"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.029395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7cf777545f-f8drb" event={"ID":"283e6982-7444-48c7-b7ac-ed4210329a08","Type":"ContainerStarted","Data":"0cdcd0fb4815f9aa1c23324306a72652651347364577112954dc427dfd596ae4"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.030859 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" event={"ID":"8b84edbe-4485-4027-bfc4-cebbb1b10543","Type":"ContainerStarted","Data":"55fc42714f1e8c2485a2f2ab2991b189f14890bb62beec8d56c4a223261fc69c"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.032289 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerStarted","Data":"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.032315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerStarted","Data":"54ae046171ebb1a5c5a1082e7114ac886653fcd28b5ecbb331be434df22b2e10"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.039690 4824 generic.go:334] "Generic (PLEG): container finished" podID="661c81f6-6854-416e-aea7-1344a635f2db" containerID="18cc2896c05b5182d3e266af8c0e8d6f107da5daa9b81a2e4d5ef8e0cf371e65" exitCode=0 Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.039760 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" event={"ID":"661c81f6-6854-416e-aea7-1344a635f2db","Type":"ContainerDied","Data":"18cc2896c05b5182d3e266af8c0e8d6f107da5daa9b81a2e4d5ef8e0cf371e65"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.039786 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" event={"ID":"661c81f6-6854-416e-aea7-1344a635f2db","Type":"ContainerStarted","Data":"e5f93c719de6d721a85fa8dea128d5a96c33c0da5f5e176c82969aadb4af41ac"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.042039 4824 generic.go:334] "Generic (PLEG): container finished" podID="f68506fe-ac80-41e7-87c0-c7b5ac8b5362" containerID="ce260d4bf12b843ead9c0a8144fa07e77b168be5d263f9833e8d46c101318a5d" exitCode=0 Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.042101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" event={"ID":"f68506fe-ac80-41e7-87c0-c7b5ac8b5362","Type":"ContainerDied","Data":"ce260d4bf12b843ead9c0a8144fa07e77b168be5d263f9833e8d46c101318a5d"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.042129 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" event={"ID":"f68506fe-ac80-41e7-87c0-c7b5ac8b5362","Type":"ContainerStarted","Data":"9f0d2bd9547e5b9af28f1727793676788462d4c9d941a44af9285262e1676154"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.042333 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dn2np" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="registry-server" containerID="cri-o://42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4" gracePeriod=2 Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.329649 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.353001 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.482932 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.483009 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.483043 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.483114 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgb7x\" (UniqueName: \"kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.483194 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.483239 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.497828 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x" (OuterVolumeSpecName: "kube-api-access-wgb7x") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "kube-api-access-wgb7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.509757 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.513536 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.521988 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.527201 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.533661 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.534984 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config" (OuterVolumeSpecName: "config") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602385 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602412 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602420 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602431 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgb7x\" (UniqueName: \"kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602439 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602447 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.706211 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9j78\" (UniqueName: \"kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78\") pod \"1d8c8cff-4340-4eb4-94b7-062968a14848\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.706317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities\") pod \"1d8c8cff-4340-4eb4-94b7-062968a14848\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.706431 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content\") pod \"1d8c8cff-4340-4eb4-94b7-062968a14848\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.708466 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities" (OuterVolumeSpecName: "utilities") pod "1d8c8cff-4340-4eb4-94b7-062968a14848" (UID: "1d8c8cff-4340-4eb4-94b7-062968a14848"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.726788 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78" (OuterVolumeSpecName: "kube-api-access-f9j78") pod "1d8c8cff-4340-4eb4-94b7-062968a14848" (UID: "1d8c8cff-4340-4eb4-94b7-062968a14848"). InnerVolumeSpecName "kube-api-access-f9j78". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.757129 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d8c8cff-4340-4eb4-94b7-062968a14848" (UID: "1d8c8cff-4340-4eb4-94b7-062968a14848"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.807896 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9j78\" (UniqueName: \"kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.807921 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.807930 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.051632 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerID="42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4" exitCode=0 Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.051891 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerDied","Data":"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.051917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerDied","Data":"2dec804f6fee46d1464f08de4ca7da8580e7f171b751ac42a7b92ec863162d9b"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.051933 4824 scope.go:117] "RemoveContainer" containerID="42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.052064 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.059406 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" event={"ID":"f68506fe-ac80-41e7-87c0-c7b5ac8b5362","Type":"ContainerDied","Data":"9f0d2bd9547e5b9af28f1727793676788462d4c9d941a44af9285262e1676154"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.059488 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.062574 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerStarted","Data":"820df2cc441bc3b9b3b9761588c4025b1f941d3a17f1376789c1aa427ba78ecd"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.063309 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.063337 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.066051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerStarted","Data":"1568d27a18b9471d6e4be5cfb1adad4950254a106a787756913dd9ddb73cf5f9"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.067488 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerStarted","Data":"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.069261 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerStarted","Data":"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.069332 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api-log" containerID="cri-o://0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" gracePeriod=30 Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.069391 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.069402 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api" containerID="cri-o://f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" gracePeriod=30 Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.079221 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" event={"ID":"661c81f6-6854-416e-aea7-1344a635f2db","Type":"ContainerStarted","Data":"4b21f7d92e0f61e8df8d11e4819f3e436ec4a36c57f8769d50008872b4efc836"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.079560 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.082171 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6b9b444cf4-q85qq" podStartSLOduration=3.082160537 podStartE2EDuration="3.082160537s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:03.079356107 +0000 UTC m=+965.372385399" watchObservedRunningTime="2026-01-21 11:27:03.082160537 +0000 UTC m=+965.375189829" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.096792 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.096779301 podStartE2EDuration="3.096779301s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:03.090422289 +0000 UTC m=+965.383451582" watchObservedRunningTime="2026-01-21 11:27:03.096779301 +0000 UTC m=+965.389808593" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.118107 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" podStartSLOduration=3.118089778 podStartE2EDuration="3.118089778s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:03.114475089 +0000 UTC m=+965.407504382" watchObservedRunningTime="2026-01-21 11:27:03.118089778 +0000 UTC m=+965.411119069" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.153787 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.161812 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.194413 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.199856 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.245712 4824 scope.go:117] "RemoveContainer" containerID="d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.493194 4824 scope.go:117] "RemoveContainer" containerID="d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.534163 4824 scope.go:117] "RemoveContainer" containerID="42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4" Jan 21 11:27:03 crc kubenswrapper[4824]: E0121 11:27:03.538785 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4\": container with ID starting with 42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4 not found: ID does not exist" containerID="42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.538819 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4"} err="failed to get container status \"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4\": rpc error: code = NotFound desc = could not find container \"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4\": container with ID starting with 42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4 not found: ID does not exist" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.538842 4824 scope.go:117] "RemoveContainer" containerID="d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943" Jan 21 11:27:03 crc kubenswrapper[4824]: E0121 11:27:03.539192 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943\": container with ID starting with d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943 not found: ID does not exist" containerID="d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.539233 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943"} err="failed to get container status \"d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943\": rpc error: code = NotFound desc = could not find container \"d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943\": container with ID starting with d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943 not found: ID does not exist" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.539261 4824 scope.go:117] "RemoveContainer" containerID="d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a" Jan 21 11:27:03 crc kubenswrapper[4824]: E0121 11:27:03.539520 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a\": container with ID starting with d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a not found: ID does not exist" containerID="d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.539560 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a"} err="failed to get container status \"d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a\": rpc error: code = NotFound desc = could not find container \"d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a\": container with ID starting with d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a not found: ID does not exist" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.539582 4824 scope.go:117] "RemoveContainer" containerID="ce260d4bf12b843ead9c0a8144fa07e77b168be5d263f9833e8d46c101318a5d" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.759582 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.932766 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.933027 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.933828 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.933142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.934053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.934359 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.934443 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.934576 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2qhr\" (UniqueName: \"kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.935205 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.934600 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs" (OuterVolumeSpecName: "logs") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.938142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.939496 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts" (OuterVolumeSpecName: "scripts") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.940010 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr" (OuterVolumeSpecName: "kube-api-access-s2qhr") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "kube-api-access-s2qhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.953411 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.981780 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data" (OuterVolumeSpecName: "config-data") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037058 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037087 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037096 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037104 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2qhr\" (UniqueName: \"kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037114 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037121 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.056470 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" path="/var/lib/kubelet/pods/1d8c8cff-4340-4eb4-94b7-062968a14848/volumes" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.057233 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f68506fe-ac80-41e7-87c0-c7b5ac8b5362" path="/var/lib/kubelet/pods/f68506fe-ac80-41e7-87c0-c7b5ac8b5362/volumes" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086435 4824 generic.go:334] "Generic (PLEG): container finished" podID="62025fca-0943-48d2-b139-c55b0c7df358" containerID="f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" exitCode=0 Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086509 4824 generic.go:334] "Generic (PLEG): container finished" podID="62025fca-0943-48d2-b139-c55b0c7df358" containerID="0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" exitCode=143 Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086590 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerDied","Data":"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerDied","Data":"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerDied","Data":"54ae046171ebb1a5c5a1082e7114ac886653fcd28b5ecbb331be434df22b2e10"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086799 4824 scope.go:117] "RemoveContainer" containerID="f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086949 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.100359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerStarted","Data":"8eb8e5c1518e1715f46644f48ab6d186888980058e318f472bef1f90eb8e3c37"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.100588 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.103838 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerStarted","Data":"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.109373 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7cf777545f-f8drb" event={"ID":"283e6982-7444-48c7-b7ac-ed4210329a08","Type":"ContainerStarted","Data":"0140e48b86398e0eb1d1abbcdeb37e528fd4aaa0df2804bbddad333abe356436"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.109478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7cf777545f-f8drb" event={"ID":"283e6982-7444-48c7-b7ac-ed4210329a08","Type":"ContainerStarted","Data":"62f66de215fca8a12184ba58b7db09842a6daa1ccadd8e7b306a2bf8dcedded7"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.121772 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.285811915 podStartE2EDuration="5.121757253s" podCreationTimestamp="2026-01-21 11:26:59 +0000 UTC" firstStartedPulling="2026-01-21 11:26:59.753143791 +0000 UTC m=+962.046173083" lastFinishedPulling="2026-01-21 11:27:03.589089129 +0000 UTC m=+965.882118421" observedRunningTime="2026-01-21 11:27:04.114808446 +0000 UTC m=+966.407837738" watchObservedRunningTime="2026-01-21 11:27:04.121757253 +0000 UTC m=+966.414786545" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.127556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" event={"ID":"8b84edbe-4485-4027-bfc4-cebbb1b10543","Type":"ContainerStarted","Data":"ab256689d77ac6ba349d939c9dfa6b029fe4bea2c7392364417c8b51ecfca90f"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.127665 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" event={"ID":"8b84edbe-4485-4027-bfc4-cebbb1b10543","Type":"ContainerStarted","Data":"450ab9f6c9cbe4618fe598d7905f883020025f5d04727d40f1f9d48dedbdd2ae"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.132550 4824 scope.go:117] "RemoveContainer" containerID="0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.147593 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.462352276 podStartE2EDuration="4.14756866s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="2026-01-21 11:27:01.308068844 +0000 UTC m=+963.601098137" lastFinishedPulling="2026-01-21 11:27:01.993285229 +0000 UTC m=+964.286314521" observedRunningTime="2026-01-21 11:27:04.132864544 +0000 UTC m=+966.425893836" watchObservedRunningTime="2026-01-21 11:27:04.14756866 +0000 UTC m=+966.440597952" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.157870 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.160523 4824 scope.go:117] "RemoveContainer" containerID="f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.165454 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434\": container with ID starting with f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434 not found: ID does not exist" containerID="f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.165491 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434"} err="failed to get container status \"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434\": rpc error: code = NotFound desc = could not find container \"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434\": container with ID starting with f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434 not found: ID does not exist" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.165512 4824 scope.go:117] "RemoveContainer" containerID="0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.165856 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9\": container with ID starting with 0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9 not found: ID does not exist" containerID="0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.165881 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9"} err="failed to get container status \"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9\": rpc error: code = NotFound desc = could not find container \"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9\": container with ID starting with 0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9 not found: ID does not exist" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.165895 4824 scope.go:117] "RemoveContainer" containerID="f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.166155 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434"} err="failed to get container status \"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434\": rpc error: code = NotFound desc = could not find container \"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434\": container with ID starting with f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434 not found: ID does not exist" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.166188 4824 scope.go:117] "RemoveContainer" containerID="0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.166399 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9"} err="failed to get container status \"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9\": rpc error: code = NotFound desc = could not find container \"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9\": container with ID starting with 0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9 not found: ID does not exist" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.167381 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.168210 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7cf777545f-f8drb" podStartSLOduration=2.03794652 podStartE2EDuration="4.168195486s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="2026-01-21 11:27:01.364393736 +0000 UTC m=+963.657423028" lastFinishedPulling="2026-01-21 11:27:03.494642702 +0000 UTC m=+965.787671994" observedRunningTime="2026-01-21 11:27:04.157287521 +0000 UTC m=+966.450316814" watchObservedRunningTime="2026-01-21 11:27:04.168195486 +0000 UTC m=+966.461224779" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179506 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179851 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api-log" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179864 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api-log" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179875 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179881 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179892 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f68506fe-ac80-41e7-87c0-c7b5ac8b5362" containerName="init" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179897 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f68506fe-ac80-41e7-87c0-c7b5ac8b5362" containerName="init" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179908 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="extract-utilities" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179914 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="extract-utilities" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179923 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="registry-server" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179928 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="registry-server" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179938 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="extract-content" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179943 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="extract-content" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.180178 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.180192 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="registry-server" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.180198 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api-log" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.180218 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f68506fe-ac80-41e7-87c0-c7b5ac8b5362" containerName="init" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.181020 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.183184 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.183331 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.183520 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.186784 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.186917 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" podStartSLOduration=2.129948567 podStartE2EDuration="4.186902442s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="2026-01-21 11:27:01.443907528 +0000 UTC m=+963.736936819" lastFinishedPulling="2026-01-21 11:27:03.500861402 +0000 UTC m=+965.793890694" observedRunningTime="2026-01-21 11:27:04.179489399 +0000 UTC m=+966.472518691" watchObservedRunningTime="2026-01-21 11:27:04.186902442 +0000 UTC m=+966.479931734" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344516 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff9de537-ecd2-469d-b90e-5857bbf64c2d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff9de537-ecd2-469d-b90e-5857bbf64c2d-logs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344667 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-scripts\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344716 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344802 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-447gr\" (UniqueName: \"kubernetes.io/projected/ff9de537-ecd2-469d-b90e-5857bbf64c2d-kube-api-access-447gr\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344819 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445764 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-scripts\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445934 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-447gr\" (UniqueName: \"kubernetes.io/projected/ff9de537-ecd2-469d-b90e-5857bbf64c2d-kube-api-access-447gr\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.446019 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.446043 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff9de537-ecd2-469d-b90e-5857bbf64c2d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.446100 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff9de537-ecd2-469d-b90e-5857bbf64c2d-logs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.446433 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff9de537-ecd2-469d-b90e-5857bbf64c2d-logs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.447071 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff9de537-ecd2-469d-b90e-5857bbf64c2d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.449190 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-scripts\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.450147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.450537 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.451526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.452163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.453602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.466619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-447gr\" (UniqueName: \"kubernetes.io/projected/ff9de537-ecd2-469d-b90e-5857bbf64c2d-kube-api-access-447gr\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.502350 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.906387 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:05 crc kubenswrapper[4824]: I0121 11:27:05.136082 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff9de537-ecd2-469d-b90e-5857bbf64c2d","Type":"ContainerStarted","Data":"67ee0780e6bb858ef355fe024566b8fb623152b9372fb6f8aad3e487374938d7"} Jan 21 11:27:05 crc kubenswrapper[4824]: I0121 11:27:05.571888 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.057588 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62025fca-0943-48d2-b139-c55b0c7df358" path="/var/lib/kubelet/pods/62025fca-0943-48d2-b139-c55b0c7df358/volumes" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.143904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff9de537-ecd2-469d-b90e-5857bbf64c2d","Type":"ContainerStarted","Data":"4300f85a676d71a079b02f932e8d9dae11e471e7fe4d4a61effa67b37578b02c"} Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.143940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff9de537-ecd2-469d-b90e-5857bbf64c2d","Type":"ContainerStarted","Data":"f2c098bd6680501c78682d7bd8ce5e2c495edc0d917df3678d84a806e3076625"} Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.163469 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.163455262 podStartE2EDuration="2.163455262s" podCreationTimestamp="2026-01-21 11:27:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:06.162747236 +0000 UTC m=+968.455776528" watchObservedRunningTime="2026-01-21 11:27:06.163455262 +0000 UTC m=+968.456484554" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.252507 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c7f965fb4-8slr2"] Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.253715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.255600 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.255902 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.265626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c7f965fb4-8slr2"] Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390078 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-public-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f7f97b5-8abd-4351-8396-cf1b76c376a7-logs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390217 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-combined-ca-bundle\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390239 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data-custom\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-internal-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f268\" (UniqueName: \"kubernetes.io/projected/3f7f97b5-8abd-4351-8396-cf1b76c376a7-kube-api-access-8f268\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491486 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-internal-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f268\" (UniqueName: \"kubernetes.io/projected/3f7f97b5-8abd-4351-8396-cf1b76c376a7-kube-api-access-8f268\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491673 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-public-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491744 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f7f97b5-8abd-4351-8396-cf1b76c376a7-logs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491805 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-combined-ca-bundle\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491837 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data-custom\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.493094 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f7f97b5-8abd-4351-8396-cf1b76c376a7-logs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.496395 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-public-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.496507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data-custom\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.496847 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-combined-ca-bundle\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.497833 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.498254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-internal-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.507535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f268\" (UniqueName: \"kubernetes.io/projected/3f7f97b5-8abd-4351-8396-cf1b76c376a7-kube-api-access-8f268\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.578303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.966534 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c7f965fb4-8slr2"] Jan 21 11:27:06 crc kubenswrapper[4824]: W0121 11:27:06.969921 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f7f97b5_8abd_4351_8396_cf1b76c376a7.slice/crio-62d120dae566220242a7fc924ec317d9e8ba45834de3906e26ea5a79327dfa28 WatchSource:0}: Error finding container 62d120dae566220242a7fc924ec317d9e8ba45834de3906e26ea5a79327dfa28: Status 404 returned error can't find the container with id 62d120dae566220242a7fc924ec317d9e8ba45834de3906e26ea5a79327dfa28 Jan 21 11:27:07 crc kubenswrapper[4824]: I0121 11:27:07.152237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7f965fb4-8slr2" event={"ID":"3f7f97b5-8abd-4351-8396-cf1b76c376a7","Type":"ContainerStarted","Data":"e7a548bf434fd3b356d458e6223ea4dc751528f6ea6cf4d9b657539003cfd16e"} Jan 21 11:27:07 crc kubenswrapper[4824]: I0121 11:27:07.152277 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7f965fb4-8slr2" event={"ID":"3f7f97b5-8abd-4351-8396-cf1b76c376a7","Type":"ContainerStarted","Data":"62d120dae566220242a7fc924ec317d9e8ba45834de3906e26ea5a79327dfa28"} Jan 21 11:27:07 crc kubenswrapper[4824]: I0121 11:27:07.152376 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 21 11:27:07 crc kubenswrapper[4824]: I0121 11:27:07.379505 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:08 crc kubenswrapper[4824]: I0121 11:27:08.161045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7f965fb4-8slr2" event={"ID":"3f7f97b5-8abd-4351-8396-cf1b76c376a7","Type":"ContainerStarted","Data":"2c342342909bc38f0c94dba0cbbc1e6ee1d8016636b9f1985115fed0b825a876"} Jan 21 11:27:08 crc kubenswrapper[4824]: I0121 11:27:08.161442 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:08 crc kubenswrapper[4824]: I0121 11:27:08.161473 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:08 crc kubenswrapper[4824]: I0121 11:27:08.184662 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c7f965fb4-8slr2" podStartSLOduration=2.184646841 podStartE2EDuration="2.184646841s" podCreationTimestamp="2026-01-21 11:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:08.179527715 +0000 UTC m=+970.472557007" watchObservedRunningTime="2026-01-21 11:27:08.184646841 +0000 UTC m=+970.477676133" Jan 21 11:27:08 crc kubenswrapper[4824]: I0121 11:27:08.670468 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:10 crc kubenswrapper[4824]: I0121 11:27:10.749337 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 21 11:27:10 crc kubenswrapper[4824]: I0121 11:27:10.789357 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.009108 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.051704 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.051884 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="dnsmasq-dns" containerID="cri-o://25b08d9ba24d1cbae3803365bca42ff0e044f0f812800389a684ef8eb1362dbe" gracePeriod=10 Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.194495 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerID="25b08d9ba24d1cbae3803365bca42ff0e044f0f812800389a684ef8eb1362dbe" exitCode=0 Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.194661 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="cinder-scheduler" containerID="cri-o://945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74" gracePeriod=30 Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.194889 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" event={"ID":"bb4ea3d2-63c7-4829-b6b8-8112476493fa","Type":"ContainerDied","Data":"25b08d9ba24d1cbae3803365bca42ff0e044f0f812800389a684ef8eb1362dbe"} Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.195119 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="probe" containerID="cri-o://b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a" gracePeriod=30 Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.600152 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678238 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678281 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clj2h\" (UniqueName: \"kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678331 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678355 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678427 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.685179 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h" (OuterVolumeSpecName: "kube-api-access-clj2h") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "kube-api-access-clj2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.715764 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.718976 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.719626 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.727252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config" (OuterVolumeSpecName: "config") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.730814 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787230 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787263 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clj2h\" (UniqueName: \"kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787275 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787283 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787291 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787299 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.862320 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.202719 4824 generic.go:334] "Generic (PLEG): container finished" podID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerID="b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a" exitCode=0 Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.202751 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerDied","Data":"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a"} Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.204744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" event={"ID":"bb4ea3d2-63c7-4829-b6b8-8112476493fa","Type":"ContainerDied","Data":"5e519680bc7c55a0ecc86ba490dcbb3969a313e5b2bc06cbb94da1a485a26d3b"} Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.204849 4824 scope.go:117] "RemoveContainer" containerID="25b08d9ba24d1cbae3803365bca42ff0e044f0f812800389a684ef8eb1362dbe" Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.205055 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.221593 4824 scope.go:117] "RemoveContainer" containerID="f5715a5c2dbadf161e906011876459ee94ba9a1523d9455c1d55ac180e574139" Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.222921 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.234492 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.597531 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617314 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617359 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617391 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617440 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmmwh\" (UniqueName: \"kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617443 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617563 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.618465 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.622126 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts" (OuterVolumeSpecName: "scripts") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.625100 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh" (OuterVolumeSpecName: "kube-api-access-cmmwh") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "kube-api-access-cmmwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.625099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.662663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.688584 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data" (OuterVolumeSpecName: "config-data") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.719863 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.719887 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmmwh\" (UniqueName: \"kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.719897 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.719907 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.719914 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.854305 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.894013 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.894170 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-79bfcdcfbd-s2v5h" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-api" containerID="cri-o://ff804ebe93733fa9e697132922f8e30c6ee75d058b7cb8afecec7da791271c35" gracePeriod=30 Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.894470 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-79bfcdcfbd-s2v5h" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-httpd" containerID="cri-o://eb6d7862490b4afbdbbe422c8984287dc94d6e4a39261b3846440fb920c542fb" gracePeriod=30 Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.070247 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" path="/var/lib/kubelet/pods/bb4ea3d2-63c7-4829-b6b8-8112476493fa/volumes" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.222788 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerID="eb6d7862490b4afbdbbe422c8984287dc94d6e4a39261b3846440fb920c542fb" exitCode=0 Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.222851 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerDied","Data":"eb6d7862490b4afbdbbe422c8984287dc94d6e4a39261b3846440fb920c542fb"} Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.224740 4824 generic.go:334] "Generic (PLEG): container finished" podID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerID="945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74" exitCode=0 Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.224785 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.224783 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerDied","Data":"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74"} Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.224922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerDied","Data":"ccfadddd012e5287b137e7bb05a0f3a68e04a10bc60fc278e3e35eab10438b65"} Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.224945 4824 scope.go:117] "RemoveContainer" containerID="b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.240914 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.243397 4824 scope.go:117] "RemoveContainer" containerID="945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.246875 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.264914 4824 scope.go:117] "RemoveContainer" containerID="b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268025 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.268378 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="probe" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268391 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="probe" Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.268415 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="init" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268420 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="init" Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.268429 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="dnsmasq-dns" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268435 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="dnsmasq-dns" Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.268448 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="cinder-scheduler" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268453 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="cinder-scheduler" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268632 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="cinder-scheduler" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268645 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="dnsmasq-dns" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268667 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="probe" Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.269163 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a\": container with ID starting with b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a not found: ID does not exist" containerID="b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.269195 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a"} err="failed to get container status \"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a\": rpc error: code = NotFound desc = could not find container \"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a\": container with ID starting with b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a not found: ID does not exist" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.269220 4824 scope.go:117] "RemoveContainer" containerID="945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.269509 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.269569 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74\": container with ID starting with 945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74 not found: ID does not exist" containerID="945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.269592 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74"} err="failed to get container status \"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74\": rpc error: code = NotFound desc = could not find container \"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74\": container with ID starting with 945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74 not found: ID does not exist" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.270975 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.278795 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.328413 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.328535 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.328710 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e0f0267-58d2-4690-9187-8ddf03ef082e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.329316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fdvv\" (UniqueName: \"kubernetes.io/projected/2e0f0267-58d2-4690-9187-8ddf03ef082e-kube-api-access-9fdvv\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.329361 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.329378 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fdvv\" (UniqueName: \"kubernetes.io/projected/2e0f0267-58d2-4690-9187-8ddf03ef082e-kube-api-access-9fdvv\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430820 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e0f0267-58d2-4690-9187-8ddf03ef082e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.431025 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e0f0267-58d2-4690-9187-8ddf03ef082e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.435144 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.435310 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.435929 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.436372 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.447849 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fdvv\" (UniqueName: \"kubernetes.io/projected/2e0f0267-58d2-4690-9187-8ddf03ef082e-kube-api-access-9fdvv\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.585492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.976505 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:14 crc kubenswrapper[4824]: W0121 11:27:14.987046 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e0f0267_58d2_4690_9187_8ddf03ef082e.slice/crio-e4b13faacb70f79ca1a74ffe6705e4d4d8ba2ae3dd39c51fa021fe0341cbc621 WatchSource:0}: Error finding container e4b13faacb70f79ca1a74ffe6705e4d4d8ba2ae3dd39c51fa021fe0341cbc621: Status 404 returned error can't find the container with id e4b13faacb70f79ca1a74ffe6705e4d4d8ba2ae3dd39c51fa021fe0341cbc621 Jan 21 11:27:15 crc kubenswrapper[4824]: I0121 11:27:15.233310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e0f0267-58d2-4690-9187-8ddf03ef082e","Type":"ContainerStarted","Data":"e4b13faacb70f79ca1a74ffe6705e4d4d8ba2ae3dd39c51fa021fe0341cbc621"} Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.064884 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.065275 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.072309 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" path="/var/lib/kubelet/pods/a5721831-a6aa-4d42-b62a-041dbe5c23ad/volumes" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.073519 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.074147 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.074198 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5" gracePeriod=600 Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.079631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.246013 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5" exitCode=0 Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.246250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5"} Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.246281 4824 scope.go:117] "RemoveContainer" containerID="0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.252409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e0f0267-58d2-4690-9187-8ddf03ef082e","Type":"ContainerStarted","Data":"17331213015ca691828e0bb69e808e23e5f81297d08aad89e5b16c716d858917"} Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.252439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e0f0267-58d2-4690-9187-8ddf03ef082e","Type":"ContainerStarted","Data":"f11ed05be21bc138f79d8f98c30bec5e9c23f6fd84a904ab2f95213a4528812d"} Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.256398 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.271833 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.271821918 podStartE2EDuration="2.271821918s" podCreationTimestamp="2026-01-21 11:27:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:16.265323351 +0000 UTC m=+978.558352643" watchObservedRunningTime="2026-01-21 11:27:16.271821918 +0000 UTC m=+978.564851209" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.380920 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.511819 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: i/o timeout" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.262159 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690"} Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.264680 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerID="ff804ebe93733fa9e697132922f8e30c6ee75d058b7cb8afecec7da791271c35" exitCode=0 Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.264778 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerDied","Data":"ff804ebe93733fa9e697132922f8e30c6ee75d058b7cb8afecec7da791271c35"} Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.264831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerDied","Data":"2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d"} Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.264843 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.276527 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.287768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs\") pod \"d9302662-76b1-4acc-b9e3-ffde925d34b2\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.287863 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config\") pod \"d9302662-76b1-4acc-b9e3-ffde925d34b2\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.287984 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x8cf\" (UniqueName: \"kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf\") pod \"d9302662-76b1-4acc-b9e3-ffde925d34b2\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.288123 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config\") pod \"d9302662-76b1-4acc-b9e3-ffde925d34b2\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.288145 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle\") pod \"d9302662-76b1-4acc-b9e3-ffde925d34b2\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.294007 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf" (OuterVolumeSpecName: "kube-api-access-9x8cf") pod "d9302662-76b1-4acc-b9e3-ffde925d34b2" (UID: "d9302662-76b1-4acc-b9e3-ffde925d34b2"). InnerVolumeSpecName "kube-api-access-9x8cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.305043 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d9302662-76b1-4acc-b9e3-ffde925d34b2" (UID: "d9302662-76b1-4acc-b9e3-ffde925d34b2"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.337057 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config" (OuterVolumeSpecName: "config") pod "d9302662-76b1-4acc-b9e3-ffde925d34b2" (UID: "d9302662-76b1-4acc-b9e3-ffde925d34b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.361914 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9302662-76b1-4acc-b9e3-ffde925d34b2" (UID: "d9302662-76b1-4acc-b9e3-ffde925d34b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.376438 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d9302662-76b1-4acc-b9e3-ffde925d34b2" (UID: "d9302662-76b1-4acc-b9e3-ffde925d34b2"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.390235 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.390262 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.390274 4824 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.390281 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.390289 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x8cf\" (UniqueName: \"kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.883367 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.982186 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.042027 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.042234 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6b9b444cf4-q85qq" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api-log" containerID="cri-o://41e3cbe62a3cfe1f1349c3e05b9d5169c6ca9dd78e1f915854df40b91a504d55" gracePeriod=30 Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.042365 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6b9b444cf4-q85qq" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api" containerID="cri-o://820df2cc441bc3b9b3b9761588c4025b1f941d3a17f1376789c1aa427ba78ecd" gracePeriod=30 Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.280611 4824 generic.go:334] "Generic (PLEG): container finished" podID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerID="41e3cbe62a3cfe1f1349c3e05b9d5169c6ca9dd78e1f915854df40b91a504d55" exitCode=143 Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.281057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerDied","Data":"41e3cbe62a3cfe1f1349c3e05b9d5169c6ca9dd78e1f915854df40b91a504d55"} Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.281149 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.299183 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.305463 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.752191 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:27:19 crc kubenswrapper[4824]: I0121 11:27:19.586181 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.057161 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" path="/var/lib/kubelet/pods/d9302662-76b1-4acc-b9e3-ffde925d34b2/volumes" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.794839 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Jan 21 11:27:20 crc kubenswrapper[4824]: E0121 11:27:20.795781 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-api" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.795873 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-api" Jan 21 11:27:20 crc kubenswrapper[4824]: E0121 11:27:20.795943 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-httpd" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.796017 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-httpd" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.796235 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-httpd" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.796292 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-api" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.796839 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.799337 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.799337 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.799388 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-mp8h2" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.805516 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.841452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config-secret\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.841495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.841780 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.841848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzwj8\" (UniqueName: \"kubernetes.io/projected/1af94d08-6d1e-4a97-8f82-27a1755d17a1-kube-api-access-tzwj8\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.943896 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.943949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzwj8\" (UniqueName: \"kubernetes.io/projected/1af94d08-6d1e-4a97-8f82-27a1755d17a1-kube-api-access-tzwj8\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.944088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config-secret\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.944119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.944858 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.948719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.954749 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config-secret\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.963247 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzwj8\" (UniqueName: \"kubernetes.io/projected/1af94d08-6d1e-4a97-8f82-27a1755d17a1-kube-api-access-tzwj8\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.114033 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.198802 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b9b444cf4-q85qq" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:44672->10.217.0.163:9311: read: connection reset by peer" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.199096 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b9b444cf4-q85qq" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:44670->10.217.0.163:9311: read: connection reset by peer" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.302460 4824 generic.go:334] "Generic (PLEG): container finished" podID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerID="820df2cc441bc3b9b3b9761588c4025b1f941d3a17f1376789c1aa427ba78ecd" exitCode=0 Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.302653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerDied","Data":"820df2cc441bc3b9b3b9761588c4025b1f941d3a17f1376789c1aa427ba78ecd"} Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.467073 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.510487 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 21 11:27:21 crc kubenswrapper[4824]: W0121 11:27:21.511259 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1af94d08_6d1e_4a97_8f82_27a1755d17a1.slice/crio-1bea49d26452dfba81b66bbb6ec52dc54c631874ef4a66db9dfc3284bdf2b640 WatchSource:0}: Error finding container 1bea49d26452dfba81b66bbb6ec52dc54c631874ef4a66db9dfc3284bdf2b640: Status 404 returned error can't find the container with id 1bea49d26452dfba81b66bbb6ec52dc54c631874ef4a66db9dfc3284bdf2b640 Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.552728 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnlw6\" (UniqueName: \"kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6\") pod \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.552794 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data\") pod \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.552820 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom\") pod \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.552884 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs\") pod \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.552911 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle\") pod \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.553286 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs" (OuterVolumeSpecName: "logs") pod "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" (UID: "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.553518 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.557510 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" (UID: "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.557901 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6" (OuterVolumeSpecName: "kube-api-access-vnlw6") pod "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" (UID: "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94"). InnerVolumeSpecName "kube-api-access-vnlw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.574598 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" (UID: "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.591682 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data" (OuterVolumeSpecName: "config-data") pod "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" (UID: "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.655172 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.655908 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnlw6\" (UniqueName: \"kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.656028 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.656086 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.318080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerDied","Data":"62531d41bed5d935b440d701f3d3a82064300927572a77364b43daf9ada74a10"} Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.318307 4824 scope.go:117] "RemoveContainer" containerID="820df2cc441bc3b9b3b9761588c4025b1f941d3a17f1376789c1aa427ba78ecd" Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.318456 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.321527 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1af94d08-6d1e-4a97-8f82-27a1755d17a1","Type":"ContainerStarted","Data":"1bea49d26452dfba81b66bbb6ec52dc54c631874ef4a66db9dfc3284bdf2b640"} Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.338780 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.344899 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.349461 4824 scope.go:117] "RemoveContainer" containerID="41e3cbe62a3cfe1f1349c3e05b9d5169c6ca9dd78e1f915854df40b91a504d55" Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.846247 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.846442 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-log" containerID="cri-o://953a68ff3f5b22fd88f3a1a2babc253f6f9d0648a9b5bca14c07b85059d64e1a" gracePeriod=30 Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.846506 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-httpd" containerID="cri-o://77d449bfcfde658bccf9c84cc9cff76978535cecd3aaeeb5e306a982392e5cf4" gracePeriod=30 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.330417 4824 generic.go:334] "Generic (PLEG): container finished" podID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerID="953a68ff3f5b22fd88f3a1a2babc253f6f9d0648a9b5bca14c07b85059d64e1a" exitCode=143 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.330473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerDied","Data":"953a68ff3f5b22fd88f3a1a2babc253f6f9d0648a9b5bca14c07b85059d64e1a"} Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.370256 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.370593 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-central-agent" containerID="cri-o://7b2e891014d0053a1d16ec02f1fc280a3a0cbb96983be7316f8030e04e57c2be" gracePeriod=30 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.370645 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="sg-core" containerID="cri-o://1568d27a18b9471d6e4be5cfb1adad4950254a106a787756913dd9ddb73cf5f9" gracePeriod=30 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.370670 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-notification-agent" containerID="cri-o://3b7c3d7bba0374d53d0ac03b9cb2de415ba41cc2525693fb788df9d2d6551fe4" gracePeriod=30 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.370702 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="proxy-httpd" containerID="cri-o://8eb8e5c1518e1715f46644f48ab6d186888980058e318f472bef1f90eb8e3c37" gracePeriod=30 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.374676 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.156:3000/\": EOF" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.062717 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" path="/var/lib/kubelet/pods/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94/volumes" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342654 4824 generic.go:334] "Generic (PLEG): container finished" podID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerID="8eb8e5c1518e1715f46644f48ab6d186888980058e318f472bef1f90eb8e3c37" exitCode=0 Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342684 4824 generic.go:334] "Generic (PLEG): container finished" podID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerID="1568d27a18b9471d6e4be5cfb1adad4950254a106a787756913dd9ddb73cf5f9" exitCode=2 Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342691 4824 generic.go:334] "Generic (PLEG): container finished" podID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerID="7b2e891014d0053a1d16ec02f1fc280a3a0cbb96983be7316f8030e04e57c2be" exitCode=0 Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342711 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerDied","Data":"8eb8e5c1518e1715f46644f48ab6d186888980058e318f472bef1f90eb8e3c37"} Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerDied","Data":"1568d27a18b9471d6e4be5cfb1adad4950254a106a787756913dd9ddb73cf5f9"} Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342754 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerDied","Data":"7b2e891014d0053a1d16ec02f1fc280a3a0cbb96983be7316f8030e04e57c2be"} Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.753398 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.967612 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4h8mx"] Jan 21 11:27:24 crc kubenswrapper[4824]: E0121 11:27:24.969269 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api-log" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.969294 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api-log" Jan 21 11:27:24 crc kubenswrapper[4824]: E0121 11:27:24.969330 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.969336 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.969504 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api-log" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.969518 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.970051 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.986671 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4h8mx"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.013344 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mwhf\" (UniqueName: \"kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.013560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.071721 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-f2xfg"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.075163 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.079940 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-f2xfg"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.089860 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-91fe-account-create-update-4xgv2"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.090657 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.092201 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.100214 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-91fe-account-create-update-4xgv2"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mwhf\" (UniqueName: \"kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115656 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgm2l\" (UniqueName: \"kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115794 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115829 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpkmt\" (UniqueName: \"kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.116122 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.130529 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mwhf\" (UniqueName: \"kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.217335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.217448 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgm2l\" (UniqueName: \"kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.217513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.217537 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpkmt\" (UniqueName: \"kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.218301 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.219772 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.235382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgm2l\" (UniqueName: \"kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.238668 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpkmt\" (UniqueName: \"kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.275143 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-dhvfz"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.283882 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.305235 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dhvfz"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.305284 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-90fb-account-create-update-m4n4l"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.305372 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.306691 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.308426 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.320063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.320146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.322935 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-90fb-account-create-update-m4n4l"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.335009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btjhr\" (UniqueName: \"kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.335115 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzjdm\" (UniqueName: \"kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.387241 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.408508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.436673 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.436749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.436876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btjhr\" (UniqueName: \"kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.436971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzjdm\" (UniqueName: \"kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.437571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.437623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.472657 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzjdm\" (UniqueName: \"kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.480617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btjhr\" (UniqueName: \"kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.485537 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-16cb-account-create-update-qjbx7"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.488608 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.490254 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.494805 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-16cb-account-create-update-qjbx7"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.538807 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frs7l\" (UniqueName: \"kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.539056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.622412 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.628920 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.632758 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-79f4c948b7-4hz4p"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.633908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.637302 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.637340 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.637528 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.640777 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.640881 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frs7l\" (UniqueName: \"kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.641388 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.643230 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-79f4c948b7-4hz4p"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.661734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frs7l\" (UniqueName: \"kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.741772 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjr4p\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-kube-api-access-tjr4p\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.741988 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-public-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-internal-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-config-data\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742292 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-combined-ca-bundle\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742351 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-run-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-etc-swift\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742400 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-log-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.839590 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843761 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-internal-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-config-data\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843845 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-combined-ca-bundle\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-run-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-etc-swift\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-log-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843917 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjr4p\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-kube-api-access-tjr4p\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843938 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-public-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.844673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-run-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.846174 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-log-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.847383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-combined-ca-bundle\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.847409 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-public-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.848548 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-internal-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.850352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-config-data\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.850437 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-etc-swift\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.860045 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjr4p\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-kube-api-access-tjr4p\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.963085 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:26 crc kubenswrapper[4824]: I0121 11:27:26.359598 4824 generic.go:334] "Generic (PLEG): container finished" podID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerID="3b7c3d7bba0374d53d0ac03b9cb2de415ba41cc2525693fb788df9d2d6551fe4" exitCode=0 Jan 21 11:27:26 crc kubenswrapper[4824]: I0121 11:27:26.359671 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerDied","Data":"3b7c3d7bba0374d53d0ac03b9cb2de415ba41cc2525693fb788df9d2d6551fe4"} Jan 21 11:27:26 crc kubenswrapper[4824]: I0121 11:27:26.362264 4824 generic.go:334] "Generic (PLEG): container finished" podID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerID="77d449bfcfde658bccf9c84cc9cff76978535cecd3aaeeb5e306a982392e5cf4" exitCode=0 Jan 21 11:27:26 crc kubenswrapper[4824]: I0121 11:27:26.362292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerDied","Data":"77d449bfcfde658bccf9c84cc9cff76978535cecd3aaeeb5e306a982392e5cf4"} Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.704410 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.806982 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807103 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807128 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807159 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807219 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmjb9\" (UniqueName: \"kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807240 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807291 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.808073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.811424 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.812762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts" (OuterVolumeSpecName: "scripts") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.815628 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9" (OuterVolumeSpecName: "kube-api-access-nmjb9") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "kube-api-access-nmjb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.817884 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.839268 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.882941 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.907991 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data" (OuterVolumeSpecName: "config-data") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908449 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908520 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908580 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908685 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908706 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908742 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908837 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908863 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8grh\" (UniqueName: \"kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909377 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909397 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909405 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909416 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmjb9\" (UniqueName: \"kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909425 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909434 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909442 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909438 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs" (OuterVolumeSpecName: "logs") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.911340 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.911949 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh" (OuterVolumeSpecName: "kube-api-access-v8grh") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "kube-api-access-v8grh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.912417 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts" (OuterVolumeSpecName: "scripts") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.942044 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.948787 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.955184 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data" (OuterVolumeSpecName: "config-data") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024258 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024290 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024304 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8grh\" (UniqueName: \"kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024313 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024339 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024385 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024395 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024404 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.039798 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-16cb-account-create-update-qjbx7"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.061348 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.065170 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4h8mx"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.126709 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.198184 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-91fe-account-create-update-4xgv2"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.205418 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dhvfz"] Jan 21 11:27:29 crc kubenswrapper[4824]: W0121 11:27:29.213176 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb24eefe0_9cb1_4672_85d8_6dfe2babebff.slice/crio-5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a WatchSource:0}: Error finding container 5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a: Status 404 returned error can't find the container with id 5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.217154 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-90fb-account-create-update-m4n4l"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.361425 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-f2xfg"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.387042 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" event={"ID":"b24eefe0-9cb1-4672-85d8-6dfe2babebff","Type":"ContainerStarted","Data":"7d5e936f6ab78e748abd4aabb374f95965e7eaaf526fd1a755b675eed79144ec"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.387086 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" event={"ID":"b24eefe0-9cb1-4672-85d8-6dfe2babebff","Type":"ContainerStarted","Data":"5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.390255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerDied","Data":"b8cc173acd912ad690d3e7aa92c24c20b537f220a873ef96f5148fc3b668366a"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.390293 4824 scope.go:117] "RemoveContainer" containerID="8eb8e5c1518e1715f46644f48ab6d186888980058e318f472bef1f90eb8e3c37" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.390344 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.392181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1af94d08-6d1e-4a97-8f82-27a1755d17a1","Type":"ContainerStarted","Data":"f8d2b72296ca177ea36b5db87f12950c56e50668e1e64af2fed5ab52cefc2cc3"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.395722 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhvfz" event={"ID":"d5162c70-9359-42d5-a02d-16140ac4e366","Type":"ContainerStarted","Data":"6fa0076da1636efc5cc0178f55ca211d393f930b6c63c3d85cac4eedd9f2d288"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.395757 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhvfz" event={"ID":"d5162c70-9359-42d5-a02d-16140ac4e366","Type":"ContainerStarted","Data":"2c6b2c2ef954db67b738c424fe09e651c130b39cd2f9702c3ab0d17616f3d362"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.401894 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" podStartSLOduration=4.401882188 podStartE2EDuration="4.401882188s" podCreationTimestamp="2026-01-21 11:27:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:29.399528649 +0000 UTC m=+991.692557940" watchObservedRunningTime="2026-01-21 11:27:29.401882188 +0000 UTC m=+991.694911481" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.401921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91fe-account-create-update-4xgv2" event={"ID":"60da7490-4aa1-4880-bfcb-e51a4ed99f91","Type":"ContainerStarted","Data":"a9bd46298904e9e7c4cccbdd1cecf51e20eaa4ef17b0e64f0ae86c7cca0ec79d"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.401981 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91fe-account-create-update-4xgv2" event={"ID":"60da7490-4aa1-4880-bfcb-e51a4ed99f91","Type":"ContainerStarted","Data":"72b068529352cd2fd0d98380f92cbe1ad287be50708214d7b3205789afcc691d"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.405793 4824 generic.go:334] "Generic (PLEG): container finished" podID="91261d07-56c8-4da3-8dd5-6f23587d77be" containerID="06b41897897bde2d466371c85738035a8ab9ee6530d1dfecdba6e71d57264bfe" exitCode=0 Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.405874 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4h8mx" event={"ID":"91261d07-56c8-4da3-8dd5-6f23587d77be","Type":"ContainerDied","Data":"06b41897897bde2d466371c85738035a8ab9ee6530d1dfecdba6e71d57264bfe"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.405911 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4h8mx" event={"ID":"91261d07-56c8-4da3-8dd5-6f23587d77be","Type":"ContainerStarted","Data":"0d0a9f9151b1e3a86a4c82524d30beed3cbaa6676c1501c4fd1a3117b7a53765"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.410352 4824 generic.go:334] "Generic (PLEG): container finished" podID="34afc8af-6ce5-4ee0-917b-326bd60d6237" containerID="7ab5f4aea14ccac0c6081bda9c7469c2c70bff8d5179b2221cca7a70044cada5" exitCode=0 Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.410421 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" event={"ID":"34afc8af-6ce5-4ee0-917b-326bd60d6237","Type":"ContainerDied","Data":"7ab5f4aea14ccac0c6081bda9c7469c2c70bff8d5179b2221cca7a70044cada5"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.410440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" event={"ID":"34afc8af-6ce5-4ee0-917b-326bd60d6237","Type":"ContainerStarted","Data":"ff32875367cf89cd527e03854db7b2ba8ad3ebac9843661fa1849f3051811312"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.416543 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerDied","Data":"5dc758f1e11243d2a910301ab89754d05c98dfa5446bb49c248a5833a44c6ce9"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.416628 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.421418 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.387191921 podStartE2EDuration="9.421400364s" podCreationTimestamp="2026-01-21 11:27:20 +0000 UTC" firstStartedPulling="2026-01-21 11:27:21.513346991 +0000 UTC m=+983.806376282" lastFinishedPulling="2026-01-21 11:27:28.547555434 +0000 UTC m=+990.840584725" observedRunningTime="2026-01-21 11:27:29.41321287 +0000 UTC m=+991.706242163" watchObservedRunningTime="2026-01-21 11:27:29.421400364 +0000 UTC m=+991.714429656" Jan 21 11:27:29 crc kubenswrapper[4824]: W0121 11:27:29.429731 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12e7657f_78fe_4b5c_b614_4449e94366b9.slice/crio-62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589 WatchSource:0}: Error finding container 62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589: Status 404 returned error can't find the container with id 62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589 Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.439438 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-79f4c948b7-4hz4p"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.443412 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-dhvfz" podStartSLOduration=4.443397575 podStartE2EDuration="4.443397575s" podCreationTimestamp="2026-01-21 11:27:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:29.428470159 +0000 UTC m=+991.721499451" watchObservedRunningTime="2026-01-21 11:27:29.443397575 +0000 UTC m=+991.736426867" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.460884 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-91fe-account-create-update-4xgv2" podStartSLOduration=4.460871335 podStartE2EDuration="4.460871335s" podCreationTimestamp="2026-01-21 11:27:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:29.442264799 +0000 UTC m=+991.735294091" watchObservedRunningTime="2026-01-21 11:27:29.460871335 +0000 UTC m=+991.753900627" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.756996 4824 scope.go:117] "RemoveContainer" containerID="1568d27a18b9471d6e4be5cfb1adad4950254a106a787756913dd9ddb73cf5f9" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.786573 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.792154 4824 scope.go:117] "RemoveContainer" containerID="3b7c3d7bba0374d53d0ac03b9cb2de415ba41cc2525693fb788df9d2d6551fe4" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.797760 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.808913 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.818070 4824 scope.go:117] "RemoveContainer" containerID="7b2e891014d0053a1d16ec02f1fc280a3a0cbb96983be7316f8030e04e57c2be" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.824998 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.841622 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.841996 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-notification-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842015 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-notification-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.842026 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="sg-core" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842032 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="sg-core" Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.842043 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-log" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842049 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-log" Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.842066 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-central-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842071 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-central-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.842084 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="proxy-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842089 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="proxy-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.842096 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842102 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842258 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-notification-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842271 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842282 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="proxy-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842310 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-log" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842321 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-central-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842333 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="sg-core" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.843709 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.846475 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.846640 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.848503 4824 scope.go:117] "RemoveContainer" containerID="77d449bfcfde658bccf9c84cc9cff76978535cecd3aaeeb5e306a982392e5cf4" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.850018 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.851322 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.853446 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.853725 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.869023 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.879080 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.895917 4824 scope.go:117] "RemoveContainer" containerID="953a68ff3f5b22fd88f3a1a2babc253f6f9d0648a9b5bca14c07b85059d64e1a" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942430 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942695 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942726 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942757 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942787 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942817 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942845 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-898h7\" (UniqueName: \"kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.943058 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.943115 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcd2n\" (UniqueName: \"kubernetes.io/projected/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-kube-api-access-pcd2n\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.943168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.943185 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.943200 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-logs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcd2n\" (UniqueName: \"kubernetes.io/projected/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-kube-api-access-pcd2n\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045187 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045206 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045225 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-logs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045284 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-898h7\" (UniqueName: \"kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045463 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045534 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-logs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.046006 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.046156 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.046154 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.050157 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.050515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.050524 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.051161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.054952 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.057912 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.063566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.065663 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-898h7\" (UniqueName: \"kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.068198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.069018 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" path="/var/lib/kubelet/pods/46f0fa4b-479a-42c5-b9b2-3f30a079be48/volumes" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.069852 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" path="/var/lib/kubelet/pods/bae23da5-948a-4883-b88d-d6ed5d6ccba7/volumes" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.074137 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcd2n\" (UniqueName: \"kubernetes.io/projected/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-kube-api-access-pcd2n\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.098199 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.171341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.176599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.426695 4824 generic.go:334] "Generic (PLEG): container finished" podID="d5162c70-9359-42d5-a02d-16140ac4e366" containerID="6fa0076da1636efc5cc0178f55ca211d393f930b6c63c3d85cac4eedd9f2d288" exitCode=0 Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.426904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhvfz" event={"ID":"d5162c70-9359-42d5-a02d-16140ac4e366","Type":"ContainerDied","Data":"6fa0076da1636efc5cc0178f55ca211d393f930b6c63c3d85cac4eedd9f2d288"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.428574 4824 generic.go:334] "Generic (PLEG): container finished" podID="60da7490-4aa1-4880-bfcb-e51a4ed99f91" containerID="a9bd46298904e9e7c4cccbdd1cecf51e20eaa4ef17b0e64f0ae86c7cca0ec79d" exitCode=0 Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.428611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91fe-account-create-update-4xgv2" event={"ID":"60da7490-4aa1-4880-bfcb-e51a4ed99f91","Type":"ContainerDied","Data":"a9bd46298904e9e7c4cccbdd1cecf51e20eaa4ef17b0e64f0ae86c7cca0ec79d"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.429867 4824 generic.go:334] "Generic (PLEG): container finished" podID="12e7657f-78fe-4b5c-b614-4449e94366b9" containerID="edd12500feb323e4dab204d7803bc5160bb50ef2b5116e5ba7686e65e4e155fc" exitCode=0 Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.429901 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f2xfg" event={"ID":"12e7657f-78fe-4b5c-b614-4449e94366b9","Type":"ContainerDied","Data":"edd12500feb323e4dab204d7803bc5160bb50ef2b5116e5ba7686e65e4e155fc"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.429914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f2xfg" event={"ID":"12e7657f-78fe-4b5c-b614-4449e94366b9","Type":"ContainerStarted","Data":"62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.442612 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-79f4c948b7-4hz4p" event={"ID":"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d","Type":"ContainerStarted","Data":"e86601c8e5d510911b89a1fe69a3df136499236541d2dc47c46a5f7da7a5581e"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.442653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-79f4c948b7-4hz4p" event={"ID":"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d","Type":"ContainerStarted","Data":"4656bc506da33d3c8ed6fc9a4ae9d4a7673549488f4d44a12d49848538d1f1ce"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.442667 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-79f4c948b7-4hz4p" event={"ID":"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d","Type":"ContainerStarted","Data":"eddc6d8deacb8d5b8460a0db4d8a825a8f5b8f3b3aa8d23bbcf51fa457d8d915"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.442704 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.442725 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.445247 4824 generic.go:334] "Generic (PLEG): container finished" podID="b24eefe0-9cb1-4672-85d8-6dfe2babebff" containerID="7d5e936f6ab78e748abd4aabb374f95965e7eaaf526fd1a755b675eed79144ec" exitCode=0 Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.445300 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" event={"ID":"b24eefe0-9cb1-4672-85d8-6dfe2babebff","Type":"ContainerDied","Data":"7d5e936f6ab78e748abd4aabb374f95965e7eaaf526fd1a755b675eed79144ec"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.501142 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-79f4c948b7-4hz4p" podStartSLOduration=5.501126765 podStartE2EDuration="5.501126765s" podCreationTimestamp="2026-01-21 11:27:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:30.493402345 +0000 UTC m=+992.786431636" watchObservedRunningTime="2026-01-21 11:27:30.501126765 +0000 UTC m=+992.794156057" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.598995 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.682545 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.799104 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.803240 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.858370 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts\") pod \"91261d07-56c8-4da3-8dd5-6f23587d77be\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.858446 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mwhf\" (UniqueName: \"kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf\") pod \"91261d07-56c8-4da3-8dd5-6f23587d77be\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.858584 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts\") pod \"34afc8af-6ce5-4ee0-917b-326bd60d6237\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.858718 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frs7l\" (UniqueName: \"kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l\") pod \"34afc8af-6ce5-4ee0-917b-326bd60d6237\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.858853 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "91261d07-56c8-4da3-8dd5-6f23587d77be" (UID: "91261d07-56c8-4da3-8dd5-6f23587d77be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.859263 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "34afc8af-6ce5-4ee0-917b-326bd60d6237" (UID: "34afc8af-6ce5-4ee0-917b-326bd60d6237"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.859463 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.859487 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.867177 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf" (OuterVolumeSpecName: "kube-api-access-6mwhf") pod "91261d07-56c8-4da3-8dd5-6f23587d77be" (UID: "91261d07-56c8-4da3-8dd5-6f23587d77be"). InnerVolumeSpecName "kube-api-access-6mwhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.867388 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l" (OuterVolumeSpecName: "kube-api-access-frs7l") pod "34afc8af-6ce5-4ee0-917b-326bd60d6237" (UID: "34afc8af-6ce5-4ee0-917b-326bd60d6237"). InnerVolumeSpecName "kube-api-access-frs7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.961666 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mwhf\" (UniqueName: \"kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.961697 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frs7l\" (UniqueName: \"kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.460567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25","Type":"ContainerStarted","Data":"fa8330abdceca8284591245f5b1c8a2c098f8d7542e6d63ae0422dce6173b423"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.460865 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25","Type":"ContainerStarted","Data":"3411927c2dfcc913f64e09b90f13938fb8a275cb1f99d46bac7986f5d1d4d172"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.462846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4h8mx" event={"ID":"91261d07-56c8-4da3-8dd5-6f23587d77be","Type":"ContainerDied","Data":"0d0a9f9151b1e3a86a4c82524d30beed3cbaa6676c1501c4fd1a3117b7a53765"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.462898 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d0a9f9151b1e3a86a4c82524d30beed3cbaa6676c1501c4fd1a3117b7a53765" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.463004 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.465245 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" event={"ID":"34afc8af-6ce5-4ee0-917b-326bd60d6237","Type":"ContainerDied","Data":"ff32875367cf89cd527e03854db7b2ba8ad3ebac9843661fa1849f3051811312"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.465292 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff32875367cf89cd527e03854db7b2ba8ad3ebac9843661fa1849f3051811312" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.465255 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.466387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerStarted","Data":"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.466417 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerStarted","Data":"24216a6df3f372a52006263ca19c690f67c9d1bbdb8896d6d5fbff2e54961470"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.774123 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.774661 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-log" containerID="cri-o://9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca" gracePeriod=30 Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.774794 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-httpd" containerID="cri-o://c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686" gracePeriod=30 Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.961288 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.988851 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.009787 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.016867 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088339 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts\") pod \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088484 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts\") pod \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088522 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts\") pod \"d5162c70-9359-42d5-a02d-16140ac4e366\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088544 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzjdm\" (UniqueName: \"kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm\") pod \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088584 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btjhr\" (UniqueName: \"kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr\") pod \"d5162c70-9359-42d5-a02d-16140ac4e366\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts\") pod \"12e7657f-78fe-4b5c-b614-4449e94366b9\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088691 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgm2l\" (UniqueName: \"kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l\") pod \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088708 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpkmt\" (UniqueName: \"kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt\") pod \"12e7657f-78fe-4b5c-b614-4449e94366b9\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.089005 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d5162c70-9359-42d5-a02d-16140ac4e366" (UID: "d5162c70-9359-42d5-a02d-16140ac4e366"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.089354 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "12e7657f-78fe-4b5c-b614-4449e94366b9" (UID: "12e7657f-78fe-4b5c-b614-4449e94366b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.089439 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.089523 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60da7490-4aa1-4880-bfcb-e51a4ed99f91" (UID: "60da7490-4aa1-4880-bfcb-e51a4ed99f91"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.089688 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b24eefe0-9cb1-4672-85d8-6dfe2babebff" (UID: "b24eefe0-9cb1-4672-85d8-6dfe2babebff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.092033 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr" (OuterVolumeSpecName: "kube-api-access-btjhr") pod "d5162c70-9359-42d5-a02d-16140ac4e366" (UID: "d5162c70-9359-42d5-a02d-16140ac4e366"). InnerVolumeSpecName "kube-api-access-btjhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.092111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt" (OuterVolumeSpecName: "kube-api-access-cpkmt") pod "12e7657f-78fe-4b5c-b614-4449e94366b9" (UID: "12e7657f-78fe-4b5c-b614-4449e94366b9"). InnerVolumeSpecName "kube-api-access-cpkmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.092194 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l" (OuterVolumeSpecName: "kube-api-access-tgm2l") pod "60da7490-4aa1-4880-bfcb-e51a4ed99f91" (UID: "60da7490-4aa1-4880-bfcb-e51a4ed99f91"). InnerVolumeSpecName "kube-api-access-tgm2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.092339 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm" (OuterVolumeSpecName: "kube-api-access-lzjdm") pod "b24eefe0-9cb1-4672-85d8-6dfe2babebff" (UID: "b24eefe0-9cb1-4672-85d8-6dfe2babebff"). InnerVolumeSpecName "kube-api-access-lzjdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191191 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191482 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzjdm\" (UniqueName: \"kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191494 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btjhr\" (UniqueName: \"kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191503 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191512 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgm2l\" (UniqueName: \"kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191520 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpkmt\" (UniqueName: \"kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191527 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.349543 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352351 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5162c70-9359-42d5-a02d-16140ac4e366" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.352456 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5162c70-9359-42d5-a02d-16140ac4e366" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352536 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60da7490-4aa1-4880-bfcb-e51a4ed99f91" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.352584 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="60da7490-4aa1-4880-bfcb-e51a4ed99f91" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352634 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24eefe0-9cb1-4672-85d8-6dfe2babebff" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.352683 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24eefe0-9cb1-4672-85d8-6dfe2babebff" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352734 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34afc8af-6ce5-4ee0-917b-326bd60d6237" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.352789 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="34afc8af-6ce5-4ee0-917b-326bd60d6237" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352844 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e7657f-78fe-4b5c-b614-4449e94366b9" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.352893 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e7657f-78fe-4b5c-b614-4449e94366b9" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352948 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91261d07-56c8-4da3-8dd5-6f23587d77be" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.353020 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="91261d07-56c8-4da3-8dd5-6f23587d77be" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.353253 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="91261d07-56c8-4da3-8dd5-6f23587d77be" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.353323 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="34afc8af-6ce5-4ee0-917b-326bd60d6237" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.353381 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e7657f-78fe-4b5c-b614-4449e94366b9" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.354110 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="60da7490-4aa1-4880-bfcb-e51a4ed99f91" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.354186 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24eefe0-9cb1-4672-85d8-6dfe2babebff" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.354251 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5162c70-9359-42d5-a02d-16140ac4e366" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.354803 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.357160 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-cj2g7" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.359033 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.376680 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.376893 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.419900 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.421618 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.429816 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.452816 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.484292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerStarted","Data":"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.487427 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25","Type":"ContainerStarted","Data":"a12c1689cb4abdc2a588fab5461b22fd986834a8bbf20ceec638f657674996c6"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.493621 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" event={"ID":"b24eefe0-9cb1-4672-85d8-6dfe2babebff","Type":"ContainerDied","Data":"5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.493722 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.493667 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.493980 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.495516 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.495831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhvfz" event={"ID":"d5162c70-9359-42d5-a02d-16140ac4e366","Type":"ContainerDied","Data":"2c6b2c2ef954db67b738c424fe09e651c130b39cd2f9702c3ab0d17616f3d362"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.495861 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c6b2c2ef954db67b738c424fe09e651c130b39cd2f9702c3ab0d17616f3d362" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.495918 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.498597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91fe-account-create-update-4xgv2" event={"ID":"60da7490-4aa1-4880-bfcb-e51a4ed99f91","Type":"ContainerDied","Data":"72b068529352cd2fd0d98380f92cbe1ad287be50708214d7b3205789afcc691d"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.498628 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72b068529352cd2fd0d98380f92cbe1ad287be50708214d7b3205789afcc691d" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.498678 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.502947 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503026 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503051 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503087 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxlch\" (UniqueName: \"kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503110 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503158 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503178 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhf6r\" (UniqueName: \"kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.507156 4824 generic.go:334] "Generic (PLEG): container finished" podID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerID="9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca" exitCode=143 Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.507229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerDied","Data":"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.511465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f2xfg" event={"ID":"12e7657f-78fe-4b5c-b614-4449e94366b9","Type":"ContainerDied","Data":"62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.511548 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.511487 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.521883 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.545555 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.550665 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.555223 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.565629 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.565619659 podStartE2EDuration="3.565619659s" podCreationTimestamp="2026-01-21 11:27:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:32.539649634 +0000 UTC m=+994.832678926" watchObservedRunningTime="2026-01-21 11:27:32.565619659 +0000 UTC m=+994.858648950" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.580673 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604739 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604798 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604827 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkztw\" (UniqueName: \"kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604866 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hljdx\" (UniqueName: \"kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604890 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604947 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605014 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605036 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605173 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxlch\" (UniqueName: \"kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605321 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605356 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605471 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605493 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhf6r\" (UniqueName: \"kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.624693 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.633565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.640619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.641012 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.641240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxlch\" (UniqueName: \"kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.644586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.646533 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhf6r\" (UniqueName: \"kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.674796 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.685371 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710015 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710056 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710114 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkztw\" (UniqueName: \"kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710191 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hljdx\" (UniqueName: \"kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710209 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710310 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710356 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710933 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.711136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.711844 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.712379 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.712571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.716679 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.716843 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.719309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.732090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hljdx\" (UniqueName: \"kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.737118 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkztw\" (UniqueName: \"kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.739930 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.906382 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.911306 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.163438 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:27:33 crc kubenswrapper[4824]: W0121 11:27:33.170199 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8d66537_20e1_4f87_9077_484e264bae81.slice/crio-6515b7c77086397376de7620cef207d703b735c2b9d2c10ad1793aedbc3eb3b1 WatchSource:0}: Error finding container 6515b7c77086397376de7620cef207d703b735c2b9d2c10ad1793aedbc3eb3b1: Status 404 returned error can't find the container with id 6515b7c77086397376de7620cef207d703b735c2b9d2c10ad1793aedbc3eb3b1 Jan 21 11:27:33 crc kubenswrapper[4824]: W0121 11:27:33.305167 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25ee21eb_c0b8_4d49_8906_4e4144abf6fc.slice/crio-3cb46ceb8d95d8862082f8003345a2758b9fe77a19473ef416fddbd583e50625 WatchSource:0}: Error finding container 3cb46ceb8d95d8862082f8003345a2758b9fe77a19473ef416fddbd583e50625: Status 404 returned error can't find the container with id 3cb46ceb8d95d8862082f8003345a2758b9fe77a19473ef416fddbd583e50625 Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.305193 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.428701 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:33 crc kubenswrapper[4824]: W0121 11:27:33.431985 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2aef2f5b_bb6b_41dd_a3eb_372e7e777122.slice/crio-986ea68f679368ee6e505b19ac9e641262bd881fd3d310847834bbf4f66d444f WatchSource:0}: Error finding container 986ea68f679368ee6e505b19ac9e641262bd881fd3d310847834bbf4f66d444f: Status 404 returned error can't find the container with id 986ea68f679368ee6e505b19ac9e641262bd881fd3d310847834bbf4f66d444f Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.445607 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:27:33 crc kubenswrapper[4824]: W0121 11:27:33.448854 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2ec0128_5881_4c67_8d78_d53d8be3a2d3.slice/crio-76a9ebfca2b2e541f5b20e22a77a9c3eda4db32f3628a468b328bf434a816a25 WatchSource:0}: Error finding container 76a9ebfca2b2e541f5b20e22a77a9c3eda4db32f3628a468b328bf434a816a25: Status 404 returned error can't find the container with id 76a9ebfca2b2e541f5b20e22a77a9c3eda4db32f3628a468b328bf434a816a25 Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.533422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5dc8c6bf78-szsb6" event={"ID":"2aef2f5b-bb6b-41dd-a3eb-372e7e777122","Type":"ContainerStarted","Data":"986ea68f679368ee6e505b19ac9e641262bd881fd3d310847834bbf4f66d444f"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.535180 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" event={"ID":"f2ec0128-5881-4c67-8d78-d53d8be3a2d3","Type":"ContainerStarted","Data":"76a9ebfca2b2e541f5b20e22a77a9c3eda4db32f3628a468b328bf434a816a25"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.536934 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerStarted","Data":"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.553092 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-656b845fb4-qst2s" event={"ID":"d8d66537-20e1-4f87-9077-484e264bae81","Type":"ContainerStarted","Data":"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.553145 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.553164 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-656b845fb4-qst2s" event={"ID":"d8d66537-20e1-4f87-9077-484e264bae81","Type":"ContainerStarted","Data":"6515b7c77086397376de7620cef207d703b735c2b9d2c10ad1793aedbc3eb3b1"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.553197 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" event={"ID":"25ee21eb-c0b8-4d49-8906-4e4144abf6fc","Type":"ContainerStarted","Data":"3cb46ceb8d95d8862082f8003345a2758b9fe77a19473ef416fddbd583e50625"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.560991 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-656b845fb4-qst2s" podStartSLOduration=1.5609775209999999 podStartE2EDuration="1.560977521s" podCreationTimestamp="2026-01-21 11:27:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:33.557485154 +0000 UTC m=+995.850514446" watchObservedRunningTime="2026-01-21 11:27:33.560977521 +0000 UTC m=+995.854006813" Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.110666 4824 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod62025fca-0943-48d2-b139-c55b0c7df358"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod62025fca-0943-48d2-b139-c55b0c7df358] : Timed out while waiting for systemd to remove kubepods-besteffort-pod62025fca_0943_48d2_b139_c55b0c7df358.slice" Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.553462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerStarted","Data":"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1"} Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.553530 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.556409 4824 generic.go:334] "Generic (PLEG): container finished" podID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerID="77516fe4a72158d203b645cac966a850920a66f17eb3fb68fcdc002cb9e2b607" exitCode=0 Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.557261 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" event={"ID":"f2ec0128-5881-4c67-8d78-d53d8be3a2d3","Type":"ContainerDied","Data":"77516fe4a72158d203b645cac966a850920a66f17eb3fb68fcdc002cb9e2b607"} Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.568744 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.178635741 podStartE2EDuration="5.568731996s" podCreationTimestamp="2026-01-21 11:27:29 +0000 UTC" firstStartedPulling="2026-01-21 11:27:30.623235249 +0000 UTC m=+992.916264541" lastFinishedPulling="2026-01-21 11:27:34.013331504 +0000 UTC m=+996.306360796" observedRunningTime="2026-01-21 11:27:34.567056976 +0000 UTC m=+996.860086268" watchObservedRunningTime="2026-01-21 11:27:34.568731996 +0000 UTC m=+996.861761288" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.541653 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.565021 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" event={"ID":"25ee21eb-c0b8-4d49-8906-4e4144abf6fc","Type":"ContainerStarted","Data":"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c"} Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.566134 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.570326 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5dc8c6bf78-szsb6" event={"ID":"2aef2f5b-bb6b-41dd-a3eb-372e7e777122","Type":"ContainerStarted","Data":"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9"} Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.570403 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.571903 4824 generic.go:334] "Generic (PLEG): container finished" podID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerID="c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686" exitCode=0 Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.571947 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.571984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerDied","Data":"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686"} Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.572006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerDied","Data":"eaaf72feb767aa6ebf839a27a99ea70de0682afc00ba0ae8e8e10dabfcaf1864"} Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.572023 4824 scope.go:117] "RemoveContainer" containerID="c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.577472 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" event={"ID":"f2ec0128-5881-4c67-8d78-d53d8be3a2d3","Type":"ContainerStarted","Data":"18927e4fff1f2c120dbfd5a267daf57c4c84dbc311e5fd297d4153df2a6f21f4"} Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.577584 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.595134 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" podStartSLOduration=1.795267887 podStartE2EDuration="3.595116211s" podCreationTimestamp="2026-01-21 11:27:32 +0000 UTC" firstStartedPulling="2026-01-21 11:27:33.307449048 +0000 UTC m=+995.600478341" lastFinishedPulling="2026-01-21 11:27:35.107297373 +0000 UTC m=+997.400326665" observedRunningTime="2026-01-21 11:27:35.584158753 +0000 UTC m=+997.877188044" watchObservedRunningTime="2026-01-21 11:27:35.595116211 +0000 UTC m=+997.888145503" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.596495 4824 scope.go:117] "RemoveContainer" containerID="9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.614560 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5dc8c6bf78-szsb6" podStartSLOduration=1.941548805 podStartE2EDuration="3.614545218s" podCreationTimestamp="2026-01-21 11:27:32 +0000 UTC" firstStartedPulling="2026-01-21 11:27:33.433808221 +0000 UTC m=+995.726837512" lastFinishedPulling="2026-01-21 11:27:35.106804634 +0000 UTC m=+997.399833925" observedRunningTime="2026-01-21 11:27:35.610138366 +0000 UTC m=+997.903167658" watchObservedRunningTime="2026-01-21 11:27:35.614545218 +0000 UTC m=+997.907574510" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.641008 4824 scope.go:117] "RemoveContainer" containerID="c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686" Jan 21 11:27:35 crc kubenswrapper[4824]: E0121 11:27:35.642537 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686\": container with ID starting with c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686 not found: ID does not exist" containerID="c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.642566 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686"} err="failed to get container status \"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686\": rpc error: code = NotFound desc = could not find container \"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686\": container with ID starting with c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686 not found: ID does not exist" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.642586 4824 scope.go:117] "RemoveContainer" containerID="9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.644793 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k6zm8"] Jan 21 11:27:35 crc kubenswrapper[4824]: E0121 11:27:35.645150 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-httpd" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645162 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-httpd" Jan 21 11:27:35 crc kubenswrapper[4824]: E0121 11:27:35.645176 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-log" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-log" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645361 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-log" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645382 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-httpd" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645466 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" podStartSLOduration=3.645457456 podStartE2EDuration="3.645457456s" podCreationTimestamp="2026-01-21 11:27:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:35.641485193 +0000 UTC m=+997.934514486" watchObservedRunningTime="2026-01-21 11:27:35.645457456 +0000 UTC m=+997.938486747" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645858 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: E0121 11:27:35.646920 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca\": container with ID starting with 9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca not found: ID does not exist" containerID="9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.646944 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca"} err="failed to get container status \"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca\": rpc error: code = NotFound desc = could not find container \"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca\": container with ID starting with 9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca not found: ID does not exist" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.647628 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gbh62" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.648276 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.648302 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.662940 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k6zm8"] Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680020 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qwlh\" (UniqueName: \"kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680160 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680208 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680230 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680267 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.685921 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts" (OuterVolumeSpecName: "scripts") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.686073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh" (OuterVolumeSpecName: "kube-api-access-7qwlh") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "kube-api-access-7qwlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.686609 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs" (OuterVolumeSpecName: "logs") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.687186 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.696059 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.751098 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.781799 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.781856 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rs4n\" (UniqueName: \"kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.781985 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782144 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782205 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782216 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qwlh\" (UniqueName: \"kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782233 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782242 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782251 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782259 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.801788 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.821683 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.840114 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data" (OuterVolumeSpecName: "config-data") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.887862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rs4n\" (UniqueName: \"kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888617 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888686 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888749 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.892925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.900087 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.900367 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.922542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rs4n\" (UniqueName: \"kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.934448 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.944422 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.955731 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.957053 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.959437 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.960149 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.974909 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.975977 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.978011 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.064023 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" path="/var/lib/kubelet/pods/12776f5e-c372-41bd-91fe-ba7d0915050a/volumes" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091213 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-scripts\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091328 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-config-data\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091429 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndtmv\" (UniqueName: \"kubernetes.io/projected/219725d0-9bef-4e42-8b78-cbc74a8ac9db-kube-api-access-ndtmv\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-logs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091546 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091562 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.138776 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.193900 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.193997 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-config-data\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.194034 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndtmv\" (UniqueName: \"kubernetes.io/projected/219725d0-9bef-4e42-8b78-cbc74a8ac9db-kube-api-access-ndtmv\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.194096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-logs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.194142 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.194152 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.197373 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.197487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-scripts\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.197615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.198835 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.199280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-logs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.201739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.204783 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-scripts\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.207004 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.204391 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-config-data\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.216933 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndtmv\" (UniqueName: \"kubernetes.io/projected/219725d0-9bef-4e42-8b78-cbc74a8ac9db-kube-api-access-ndtmv\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.225911 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.275660 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.538140 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k6zm8"] Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.586803 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" event={"ID":"9c73eda8-d376-4e18-9b33-f5ac4d10d091","Type":"ContainerStarted","Data":"bec80828734b592cc51fee125cf162b17f1a9431a491c60a24c542b64ae9e046"} Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.716408 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:37 crc kubenswrapper[4824]: I0121 11:27:37.595984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"219725d0-9bef-4e42-8b78-cbc74a8ac9db","Type":"ContainerStarted","Data":"7445af8ee234b9314f9199fd51a796bd220ddda47357119a6f2a23774de972d8"} Jan 21 11:27:37 crc kubenswrapper[4824]: I0121 11:27:37.596439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"219725d0-9bef-4e42-8b78-cbc74a8ac9db","Type":"ContainerStarted","Data":"6ba7b3bf3ee8993c98598c253f5273552426d07f3b2484dda5a710dacf8f30ac"} Jan 21 11:27:37 crc kubenswrapper[4824]: I0121 11:27:37.596454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"219725d0-9bef-4e42-8b78-cbc74a8ac9db","Type":"ContainerStarted","Data":"3e89604df6712be074b7a1e1d8001bbe51252f5826bd96b032e5c1523e48ea5d"} Jan 21 11:27:37 crc kubenswrapper[4824]: I0121 11:27:37.611319 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.611303981 podStartE2EDuration="2.611303981s" podCreationTimestamp="2026-01-21 11:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:37.609299842 +0000 UTC m=+999.902329134" watchObservedRunningTime="2026-01-21 11:27:37.611303981 +0000 UTC m=+999.904333274" Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.166890 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.167138 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-central-agent" containerID="cri-o://02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" gracePeriod=30 Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.167179 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="proxy-httpd" containerID="cri-o://184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" gracePeriod=30 Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.167205 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="sg-core" containerID="cri-o://af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" gracePeriod=30 Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.167197 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-notification-agent" containerID="cri-o://01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" gracePeriod=30 Jan 21 11:27:38 crc kubenswrapper[4824]: E0121 11:27:38.424602 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda89ff227_6009_4377_8f9f_34ac4078646f.slice/crio-184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda89ff227_6009_4377_8f9f_34ac4078646f.slice/crio-conmon-184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1.scope\": RecentStats: unable to find data in memory cache]" Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.606472 4824 generic.go:334] "Generic (PLEG): container finished" podID="a89ff227-6009-4377-8f9f-34ac4078646f" containerID="184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" exitCode=0 Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.606501 4824 generic.go:334] "Generic (PLEG): container finished" podID="a89ff227-6009-4377-8f9f-34ac4078646f" containerID="af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" exitCode=2 Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.606554 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerDied","Data":"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1"} Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.606591 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerDied","Data":"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca"} Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.016298 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.067857 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.067897 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.067914 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.068034 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.068064 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-898h7\" (UniqueName: \"kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.068097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.068153 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.069116 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.069147 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.072628 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts" (OuterVolumeSpecName: "scripts") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.088549 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7" (OuterVolumeSpecName: "kube-api-access-898h7") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "kube-api-access-898h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.100739 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.128615 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.143437 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data" (OuterVolumeSpecName: "config-data") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170739 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170797 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-898h7\" (UniqueName: \"kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170808 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170816 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170824 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170831 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170840 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.366584 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5bc6fc99dd-nv45v"] Jan 21 11:27:39 crc kubenswrapper[4824]: E0121 11:27:39.366903 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-central-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.366921 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-central-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: E0121 11:27:39.366940 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="sg-core" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.366946 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="sg-core" Jan 21 11:27:39 crc kubenswrapper[4824]: E0121 11:27:39.366986 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-notification-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.366992 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-notification-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: E0121 11:27:39.367004 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="proxy-httpd" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367009 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="proxy-httpd" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367151 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="proxy-httpd" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367163 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="sg-core" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367173 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-notification-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367188 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-central-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367662 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.382660 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5bc6fc99dd-nv45v"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.401123 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.402180 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.421003 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.427575 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.451377 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.462265 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474785 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474832 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474891 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsb2r\" (UniqueName: \"kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474912 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474928 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474980 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data-custom\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.475053 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4787\" (UniqueName: \"kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.475075 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.475154 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfwhs\" (UniqueName: \"kubernetes.io/projected/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-kube-api-access-bfwhs\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.475169 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-combined-ca-bundle\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.475185 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576585 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfwhs\" (UniqueName: \"kubernetes.io/projected/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-kube-api-access-bfwhs\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576632 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-combined-ca-bundle\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576699 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576726 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576799 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsb2r\" (UniqueName: \"kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576820 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data-custom\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576951 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4787\" (UniqueName: \"kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.580841 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data-custom\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.581549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.582075 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.582142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.585603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.588565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.588658 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-combined-ca-bundle\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.589100 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.589370 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.590043 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsb2r\" (UniqueName: \"kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.590511 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfwhs\" (UniqueName: \"kubernetes.io/projected/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-kube-api-access-bfwhs\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.590665 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4787\" (UniqueName: \"kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615894 4824 generic.go:334] "Generic (PLEG): container finished" podID="a89ff227-6009-4377-8f9f-34ac4078646f" containerID="01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" exitCode=0 Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615919 4824 generic.go:334] "Generic (PLEG): container finished" podID="a89ff227-6009-4377-8f9f-34ac4078646f" containerID="02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" exitCode=0 Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615936 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerDied","Data":"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836"} Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615947 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615971 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerDied","Data":"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2"} Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615982 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerDied","Data":"24216a6df3f372a52006263ca19c690f67c9d1bbdb8896d6d5fbff2e54961470"} Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615996 4824 scope.go:117] "RemoveContainer" containerID="184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.668737 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.681856 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.686933 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.697483 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.718933 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.718948 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.722829 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.723007 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.723417 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.748834 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-928nq\" (UniqueName: \"kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780607 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780721 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780766 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-928nq\" (UniqueName: \"kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882537 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882554 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882653 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.883535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.883742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.888231 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.889359 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.890836 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.894375 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.895802 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-928nq\" (UniqueName: \"kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.048336 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.058944 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" path="/var/lib/kubelet/pods/a89ff227-6009-4377-8f9f-34ac4078646f/volumes" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.177508 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.177555 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.207467 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.228273 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.444824 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.445075 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" containerID="cri-o://a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c" gracePeriod=60 Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.453624 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.453949 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-5dc8c6bf78-szsb6" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" containerID="cri-o://494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9" gracePeriod=60 Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.457351 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.178:8000/healthcheck\": EOF" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.462636 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5dc8c6bf78-szsb6" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.180:8004/healthcheck\": EOF" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.468234 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-68b8cf894c-ws5zd"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.469168 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.471542 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.471678 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.492951 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-68b8cf894c-ws5zd"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.505022 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-549445f5fc-fwxlz"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.506132 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.508133 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.508133 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.526445 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-549445f5fc-fwxlz"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.595910 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-combined-ca-bundle\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.595975 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-public-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.595996 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-internal-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596173 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data-custom\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596198 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpkl5\" (UniqueName: \"kubernetes.io/projected/a1aefb21-aac8-4f8f-a402-16006c1f336a-kube-api-access-zpkl5\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596219 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data-custom\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5sg2\" (UniqueName: \"kubernetes.io/projected/8655cad5-2680-4cf6-b50a-0745adeeb469-kube-api-access-z5sg2\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596289 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-combined-ca-bundle\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596322 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-internal-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-public-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.625966 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.625995 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697566 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data-custom\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpkl5\" (UniqueName: \"kubernetes.io/projected/a1aefb21-aac8-4f8f-a402-16006c1f336a-kube-api-access-zpkl5\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697629 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data-custom\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697674 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5sg2\" (UniqueName: \"kubernetes.io/projected/8655cad5-2680-4cf6-b50a-0745adeeb469-kube-api-access-z5sg2\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-combined-ca-bundle\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697770 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-internal-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697815 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-public-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697901 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-combined-ca-bundle\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.698454 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-public-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.699050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-internal-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.702552 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-combined-ca-bundle\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.707995 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data-custom\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.708093 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-combined-ca-bundle\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.708479 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.708921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-public-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.709118 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data-custom\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.709531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-internal-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.712215 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-internal-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.717411 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-public-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.720313 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5sg2\" (UniqueName: \"kubernetes.io/projected/8655cad5-2680-4cf6-b50a-0745adeeb469-kube-api-access-z5sg2\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.724334 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpkl5\" (UniqueName: \"kubernetes.io/projected/a1aefb21-aac8-4f8f-a402-16006c1f336a-kube-api-access-zpkl5\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.724586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.784522 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.823258 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.333609 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.393560 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.685605 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.908933 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.954882 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.955086 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="dnsmasq-dns" containerID="cri-o://4b21f7d92e0f61e8df8d11e4819f3e436ec4a36c57f8769d50008872b4efc836" gracePeriod=10 Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.651924 4824 generic.go:334] "Generic (PLEG): container finished" podID="661c81f6-6854-416e-aea7-1344a635f2db" containerID="4b21f7d92e0f61e8df8d11e4819f3e436ec4a36c57f8769d50008872b4efc836" exitCode=0 Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.652872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" event={"ID":"661c81f6-6854-416e-aea7-1344a635f2db","Type":"ContainerDied","Data":"4b21f7d92e0f61e8df8d11e4819f3e436ec4a36c57f8769d50008872b4efc836"} Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.731240 4824 scope.go:117] "RemoveContainer" containerID="af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.770586 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868171 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4ttl\" (UniqueName: \"kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868488 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868647 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868680 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868705 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.886419 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl" (OuterVolumeSpecName: "kube-api-access-s4ttl") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "kube-api-access-s4ttl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.895223 4824 scope.go:117] "RemoveContainer" containerID="01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.927686 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.935276 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.942252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.946397 4824 scope.go:117] "RemoveContainer" containerID="02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.950297 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.955572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config" (OuterVolumeSpecName: "config") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970792 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970815 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970824 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970834 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4ttl\" (UniqueName: \"kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970842 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970849 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.983080 4824 scope.go:117] "RemoveContainer" containerID="184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" Jan 21 11:27:43 crc kubenswrapper[4824]: E0121 11:27:43.983472 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1\": container with ID starting with 184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1 not found: ID does not exist" containerID="184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.983507 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1"} err="failed to get container status \"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1\": rpc error: code = NotFound desc = could not find container \"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1\": container with ID starting with 184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1 not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.983532 4824 scope.go:117] "RemoveContainer" containerID="af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" Jan 21 11:27:43 crc kubenswrapper[4824]: E0121 11:27:43.985631 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca\": container with ID starting with af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca not found: ID does not exist" containerID="af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.985660 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca"} err="failed to get container status \"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca\": rpc error: code = NotFound desc = could not find container \"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca\": container with ID starting with af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.985681 4824 scope.go:117] "RemoveContainer" containerID="01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" Jan 21 11:27:43 crc kubenswrapper[4824]: E0121 11:27:43.986072 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836\": container with ID starting with 01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836 not found: ID does not exist" containerID="01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.986120 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836"} err="failed to get container status \"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836\": rpc error: code = NotFound desc = could not find container \"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836\": container with ID starting with 01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836 not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.986143 4824 scope.go:117] "RemoveContainer" containerID="02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" Jan 21 11:27:43 crc kubenswrapper[4824]: E0121 11:27:43.986913 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2\": container with ID starting with 02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2 not found: ID does not exist" containerID="02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.986939 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2"} err="failed to get container status \"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2\": rpc error: code = NotFound desc = could not find container \"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2\": container with ID starting with 02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2 not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.986975 4824 scope.go:117] "RemoveContainer" containerID="184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.988102 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1"} err="failed to get container status \"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1\": rpc error: code = NotFound desc = could not find container \"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1\": container with ID starting with 184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1 not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.988127 4824 scope.go:117] "RemoveContainer" containerID="af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.988614 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca"} err="failed to get container status \"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca\": rpc error: code = NotFound desc = could not find container \"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca\": container with ID starting with af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.988643 4824 scope.go:117] "RemoveContainer" containerID="01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.989065 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836"} err="failed to get container status \"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836\": rpc error: code = NotFound desc = could not find container \"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836\": container with ID starting with 01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836 not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.989087 4824 scope.go:117] "RemoveContainer" containerID="02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.989426 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2"} err="failed to get container status \"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2\": rpc error: code = NotFound desc = could not find container \"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2\": container with ID starting with 02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2 not found: ID does not exist" Jan 21 11:27:44 crc kubenswrapper[4824]: W0121 11:27:44.300853 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8655cad5_2680_4cf6_b50a_0745adeeb469.slice/crio-a44e730098ab7c3cdd3ba90cf4c5cbf78ade4baaed854eb0ab5682d139719e48 WatchSource:0}: Error finding container a44e730098ab7c3cdd3ba90cf4c5cbf78ade4baaed854eb0ab5682d139719e48: Status 404 returned error can't find the container with id a44e730098ab7c3cdd3ba90cf4c5cbf78ade4baaed854eb0ab5682d139719e48 Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.302095 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-549445f5fc-fwxlz"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.312626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.440981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.448586 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5bc6fc99dd-nv45v"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.456044 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-68b8cf894c-ws5zd"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.468886 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.662322 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5bc6fc99dd-nv45v" event={"ID":"ca8530d7-bcfb-45fa-b0ff-a2702c081bda","Type":"ContainerStarted","Data":"5e735e3d39be9fd0c79178ef7f036c4cf4579854605131e05065dea5855f0562"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.664164 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b744b9d97-zkcv4" event={"ID":"f5046792-741c-487c-adb8-43502e9fdba1","Type":"ContainerStarted","Data":"8b94a9402348e00a388234d7d2563fa253b97825d6e2edc2c72d291efaccf943"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.664683 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.666944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" event={"ID":"a1aefb21-aac8-4f8f-a402-16006c1f336a","Type":"ContainerStarted","Data":"d74e4c157d87bd34241fbac801d56eecd0141e60287193222fe70e309b760b68"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.670800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" event={"ID":"661c81f6-6854-416e-aea7-1344a635f2db","Type":"ContainerDied","Data":"e5f93c719de6d721a85fa8dea128d5a96c33c0da5f5e176c82969aadb4af41ac"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.670821 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.670832 4824 scope.go:117] "RemoveContainer" containerID="4b21f7d92e0f61e8df8d11e4819f3e436ec4a36c57f8769d50008872b4efc836" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.672069 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-549445f5fc-fwxlz" event={"ID":"8655cad5-2680-4cf6-b50a-0745adeeb469","Type":"ContainerStarted","Data":"a44e730098ab7c3cdd3ba90cf4c5cbf78ade4baaed854eb0ab5682d139719e48"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.679996 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7b744b9d97-zkcv4" podStartSLOduration=5.679984606 podStartE2EDuration="5.679984606s" podCreationTimestamp="2026-01-21 11:27:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:44.674308249 +0000 UTC m=+1006.967337541" watchObservedRunningTime="2026-01-21 11:27:44.679984606 +0000 UTC m=+1006.973013898" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.680312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" event={"ID":"9c73eda8-d376-4e18-9b33-f5ac4d10d091","Type":"ContainerStarted","Data":"c2bb2e93e558c6a9e9ce5d9ae43e9812275d36c293fd644336fd7e9fcc3ba6c0"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.686287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerStarted","Data":"cd894bcfd6b429a651ad2e896089b4e35850d9690bc597b40bbe1fa904affe57"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.689668 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" event={"ID":"305dccfd-067c-4062-b9d1-0200e105e8fc","Type":"ContainerStarted","Data":"88223e1727fde2b87a86e2d66dfde6e4b257cc1ef3f6ba548c34ddd1445f80ea"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.707644 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.715488 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.719471 4824 scope.go:117] "RemoveContainer" containerID="18cc2896c05b5182d3e266af8c0e8d6f107da5daa9b81a2e4d5ef8e0cf371e65" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.720026 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" podStartSLOduration=2.366957178 podStartE2EDuration="9.720011316s" podCreationTimestamp="2026-01-21 11:27:35 +0000 UTC" firstStartedPulling="2026-01-21 11:27:36.542165652 +0000 UTC m=+998.835194944" lastFinishedPulling="2026-01-21 11:27:43.895219791 +0000 UTC m=+1006.188249082" observedRunningTime="2026-01-21 11:27:44.697166845 +0000 UTC m=+1006.990196137" watchObservedRunningTime="2026-01-21 11:27:44.720011316 +0000 UTC m=+1007.013040608" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.843180 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.178:8000/healthcheck\": read tcp 10.217.0.2:58164->10.217.0.178:8000: read: connection reset by peer" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.847374 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5dc8c6bf78-szsb6" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.180:8004/healthcheck\": read tcp 10.217.0.2:48836->10.217.0.180:8004: read: connection reset by peer" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.140879 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.197467 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkztw\" (UniqueName: \"kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw\") pod \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.197814 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle\") pod \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.198078 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data\") pod \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.198200 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom\") pod \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.204044 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2aef2f5b-bb6b-41dd-a3eb-372e7e777122" (UID: "2aef2f5b-bb6b-41dd-a3eb-372e7e777122"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.222078 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw" (OuterVolumeSpecName: "kube-api-access-dkztw") pod "2aef2f5b-bb6b-41dd-a3eb-372e7e777122" (UID: "2aef2f5b-bb6b-41dd-a3eb-372e7e777122"). InnerVolumeSpecName "kube-api-access-dkztw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.271945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2aef2f5b-bb6b-41dd-a3eb-372e7e777122" (UID: "2aef2f5b-bb6b-41dd-a3eb-372e7e777122"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.279225 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data" (OuterVolumeSpecName: "config-data") pod "2aef2f5b-bb6b-41dd-a3eb-372e7e777122" (UID: "2aef2f5b-bb6b-41dd-a3eb-372e7e777122"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.301858 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.301967 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.302038 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkztw\" (UniqueName: \"kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.302098 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.310470 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.403214 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom\") pod \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.403334 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data\") pod \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.403444 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxlch\" (UniqueName: \"kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch\") pod \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.403472 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle\") pod \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.406305 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "25ee21eb-c0b8-4d49-8906-4e4144abf6fc" (UID: "25ee21eb-c0b8-4d49-8906-4e4144abf6fc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.407167 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch" (OuterVolumeSpecName: "kube-api-access-kxlch") pod "25ee21eb-c0b8-4d49-8906-4e4144abf6fc" (UID: "25ee21eb-c0b8-4d49-8906-4e4144abf6fc"). InnerVolumeSpecName "kube-api-access-kxlch". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.427233 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25ee21eb-c0b8-4d49-8906-4e4144abf6fc" (UID: "25ee21eb-c0b8-4d49-8906-4e4144abf6fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.442584 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data" (OuterVolumeSpecName: "config-data") pod "25ee21eb-c0b8-4d49-8906-4e4144abf6fc" (UID: "25ee21eb-c0b8-4d49-8906-4e4144abf6fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.505613 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxlch\" (UniqueName: \"kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.505738 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.505808 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.505859 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.711610 4824 generic.go:334] "Generic (PLEG): container finished" podID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerID="494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9" exitCode=0 Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.712598 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5dc8c6bf78-szsb6" event={"ID":"2aef2f5b-bb6b-41dd-a3eb-372e7e777122","Type":"ContainerDied","Data":"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.712705 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5dc8c6bf78-szsb6" event={"ID":"2aef2f5b-bb6b-41dd-a3eb-372e7e777122","Type":"ContainerDied","Data":"986ea68f679368ee6e505b19ac9e641262bd881fd3d310847834bbf4f66d444f"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.712800 4824 scope.go:117] "RemoveContainer" containerID="494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.712965 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.722940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-549445f5fc-fwxlz" event={"ID":"8655cad5-2680-4cf6-b50a-0745adeeb469","Type":"ContainerStarted","Data":"7ca3aa4524f443c234ea4b2cf034ce17a404f31b8799b420c3731316971fe010"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.723863 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.725169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5bc6fc99dd-nv45v" event={"ID":"ca8530d7-bcfb-45fa-b0ff-a2702c081bda","Type":"ContainerStarted","Data":"cb9f39deb21448da8b4f6bd8fbe7c0a85276c9c2b132f84f1777f124f04d67b2"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.725785 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.731559 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerStarted","Data":"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.745188 4824 generic.go:334] "Generic (PLEG): container finished" podID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerID="a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c" exitCode=0 Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.745303 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" event={"ID":"25ee21eb-c0b8-4d49-8906-4e4144abf6fc","Type":"ContainerDied","Data":"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.745390 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" event={"ID":"25ee21eb-c0b8-4d49-8906-4e4144abf6fc","Type":"ContainerDied","Data":"3cb46ceb8d95d8862082f8003345a2758b9fe77a19473ef416fddbd583e50625"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.745490 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.772181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" event={"ID":"a1aefb21-aac8-4f8f-a402-16006c1f336a","Type":"ContainerStarted","Data":"54cc72cdba4177c799b1f21f80543fa4853b99e45fa1181b099b6665e7615077"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.772862 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.774351 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-549445f5fc-fwxlz" podStartSLOduration=5.774335331 podStartE2EDuration="5.774335331s" podCreationTimestamp="2026-01-21 11:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:45.760189208 +0000 UTC m=+1008.053218500" watchObservedRunningTime="2026-01-21 11:27:45.774335331 +0000 UTC m=+1008.067364623" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.777056 4824 scope.go:117] "RemoveContainer" containerID="494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9" Jan 21 11:27:45 crc kubenswrapper[4824]: E0121 11:27:45.779132 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9\": container with ID starting with 494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9 not found: ID does not exist" containerID="494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.779162 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9"} err="failed to get container status \"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9\": rpc error: code = NotFound desc = could not find container \"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9\": container with ID starting with 494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9 not found: ID does not exist" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.779182 4824 scope.go:117] "RemoveContainer" containerID="a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.782925 4824 generic.go:334] "Generic (PLEG): container finished" podID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerID="439d46cbb2fda2e5e73bdff7822009c565ff5eca78f542d8aa2a37a671b5451c" exitCode=1 Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.783028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" event={"ID":"305dccfd-067c-4062-b9d1-0200e105e8fc","Type":"ContainerDied","Data":"439d46cbb2fda2e5e73bdff7822009c565ff5eca78f542d8aa2a37a671b5451c"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.783610 4824 scope.go:117] "RemoveContainer" containerID="439d46cbb2fda2e5e73bdff7822009c565ff5eca78f542d8aa2a37a671b5451c" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.790388 4824 generic.go:334] "Generic (PLEG): container finished" podID="f5046792-741c-487c-adb8-43502e9fdba1" containerID="8eff8df136d30e860308b8a313f1f2d3a512dae3641e1576295eeeb32d6ca5d6" exitCode=1 Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.790423 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b744b9d97-zkcv4" event={"ID":"f5046792-741c-487c-adb8-43502e9fdba1","Type":"ContainerDied","Data":"8eff8df136d30e860308b8a313f1f2d3a512dae3641e1576295eeeb32d6ca5d6"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.790825 4824 scope.go:117] "RemoveContainer" containerID="8eff8df136d30e860308b8a313f1f2d3a512dae3641e1576295eeeb32d6ca5d6" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.793487 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.809166 4824 scope.go:117] "RemoveContainer" containerID="a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c" Jan 21 11:27:45 crc kubenswrapper[4824]: E0121 11:27:45.811105 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c\": container with ID starting with a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c not found: ID does not exist" containerID="a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.811139 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c"} err="failed to get container status \"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c\": rpc error: code = NotFound desc = could not find container \"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c\": container with ID starting with a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c not found: ID does not exist" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.822047 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.829560 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5bc6fc99dd-nv45v" podStartSLOduration=6.829544829 podStartE2EDuration="6.829544829s" podCreationTimestamp="2026-01-21 11:27:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:45.797452105 +0000 UTC m=+1008.090481397" watchObservedRunningTime="2026-01-21 11:27:45.829544829 +0000 UTC m=+1008.122574121" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.839381 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" podStartSLOduration=5.839370202 podStartE2EDuration="5.839370202s" podCreationTimestamp="2026-01-21 11:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:45.814554643 +0000 UTC m=+1008.107583936" watchObservedRunningTime="2026-01-21 11:27:45.839370202 +0000 UTC m=+1008.132399494" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.849911 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.857155 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.057843 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" path="/var/lib/kubelet/pods/25ee21eb-c0b8-4d49-8906-4e4144abf6fc/volumes" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.058431 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" path="/var/lib/kubelet/pods/2aef2f5b-bb6b-41dd-a3eb-372e7e777122/volumes" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.058908 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="661c81f6-6854-416e-aea7-1344a635f2db" path="/var/lib/kubelet/pods/661c81f6-6854-416e-aea7-1344a635f2db/volumes" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.277369 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.277430 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.300749 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.311739 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.799900 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerStarted","Data":"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2"} Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.800163 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerStarted","Data":"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494"} Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.803661 4824 generic.go:334] "Generic (PLEG): container finished" podID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerID="215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862" exitCode=1 Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.803728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" event={"ID":"305dccfd-067c-4062-b9d1-0200e105e8fc","Type":"ContainerDied","Data":"215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862"} Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.803773 4824 scope.go:117] "RemoveContainer" containerID="439d46cbb2fda2e5e73bdff7822009c565ff5eca78f542d8aa2a37a671b5451c" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.804211 4824 scope.go:117] "RemoveContainer" containerID="215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862" Jan 21 11:27:46 crc kubenswrapper[4824]: E0121 11:27:46.804483 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58f47bdb5-gk4p7_openstack(305dccfd-067c-4062-b9d1-0200e105e8fc)\"" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.813454 4824 generic.go:334] "Generic (PLEG): container finished" podID="f5046792-741c-487c-adb8-43502e9fdba1" containerID="e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc" exitCode=1 Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.822784 4824 scope.go:117] "RemoveContainer" containerID="e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc" Jan 21 11:27:46 crc kubenswrapper[4824]: E0121 11:27:46.823093 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7b744b9d97-zkcv4_openstack(f5046792-741c-487c-adb8-43502e9fdba1)\"" pod="openstack/heat-api-7b744b9d97-zkcv4" podUID="f5046792-741c-487c-adb8-43502e9fdba1" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.824085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b744b9d97-zkcv4" event={"ID":"f5046792-741c-487c-adb8-43502e9fdba1","Type":"ContainerDied","Data":"e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc"} Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.824119 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.824266 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.869948 4824 scope.go:117] "RemoveContainer" containerID="8eff8df136d30e860308b8a313f1f2d3a512dae3641e1576295eeeb32d6ca5d6" Jan 21 11:27:47 crc kubenswrapper[4824]: I0121 11:27:47.822949 4824 scope.go:117] "RemoveContainer" containerID="e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc" Jan 21 11:27:47 crc kubenswrapper[4824]: I0121 11:27:47.824164 4824 scope.go:117] "RemoveContainer" containerID="215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862" Jan 21 11:27:47 crc kubenswrapper[4824]: E0121 11:27:47.824360 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58f47bdb5-gk4p7_openstack(305dccfd-067c-4062-b9d1-0200e105e8fc)\"" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" Jan 21 11:27:47 crc kubenswrapper[4824]: E0121 11:27:47.824642 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7b744b9d97-zkcv4_openstack(f5046792-741c-487c-adb8-43502e9fdba1)\"" pod="openstack/heat-api-7b744b9d97-zkcv4" podUID="f5046792-741c-487c-adb8-43502e9fdba1" Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.455883 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.462255 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.832867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerStarted","Data":"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6"} Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.833137 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-central-agent" containerID="cri-o://1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc" gracePeriod=30 Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.833223 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="proxy-httpd" containerID="cri-o://526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6" gracePeriod=30 Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.833265 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="sg-core" containerID="cri-o://046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2" gracePeriod=30 Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.833294 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-notification-agent" containerID="cri-o://2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494" gracePeriod=30 Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.851998 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.112202346 podStartE2EDuration="9.851951037s" podCreationTimestamp="2026-01-21 11:27:39 +0000 UTC" firstStartedPulling="2026-01-21 11:27:44.470505931 +0000 UTC m=+1006.763535223" lastFinishedPulling="2026-01-21 11:27:48.210254621 +0000 UTC m=+1010.503283914" observedRunningTime="2026-01-21 11:27:48.847988472 +0000 UTC m=+1011.141017765" watchObservedRunningTime="2026-01-21 11:27:48.851951037 +0000 UTC m=+1011.144980328" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.724077 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.724302 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.724912 4824 scope.go:117] "RemoveContainer" containerID="215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862" Jan 21 11:27:49 crc kubenswrapper[4824]: E0121 11:27:49.725257 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58f47bdb5-gk4p7_openstack(305dccfd-067c-4062-b9d1-0200e105e8fc)\"" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.749618 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.749666 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.750280 4824 scope.go:117] "RemoveContainer" containerID="e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc" Jan 21 11:27:49 crc kubenswrapper[4824]: E0121 11:27:49.750533 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7b744b9d97-zkcv4_openstack(f5046792-741c-487c-adb8-43502e9fdba1)\"" pod="openstack/heat-api-7b744b9d97-zkcv4" podUID="f5046792-741c-487c-adb8-43502e9fdba1" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842722 4824 generic.go:334] "Generic (PLEG): container finished" podID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerID="526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6" exitCode=0 Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842754 4824 generic.go:334] "Generic (PLEG): container finished" podID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerID="046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2" exitCode=2 Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842793 4824 generic.go:334] "Generic (PLEG): container finished" podID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerID="2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494" exitCode=0 Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerDied","Data":"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6"} Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerDied","Data":"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2"} Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerDied","Data":"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494"} Jan 21 11:27:50 crc kubenswrapper[4824]: I0121 11:27:50.850861 4824 generic.go:334] "Generic (PLEG): container finished" podID="9c73eda8-d376-4e18-9b33-f5ac4d10d091" containerID="c2bb2e93e558c6a9e9ce5d9ae43e9812275d36c293fd644336fd7e9fcc3ba6c0" exitCode=0 Jan 21 11:27:50 crc kubenswrapper[4824]: I0121 11:27:50.850905 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" event={"ID":"9c73eda8-d376-4e18-9b33-f5ac4d10d091","Type":"ContainerDied","Data":"c2bb2e93e558c6a9e9ce5d9ae43e9812275d36c293fd644336fd7e9fcc3ba6c0"} Jan 21 11:27:51 crc kubenswrapper[4824]: I0121 11:27:51.904290 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:51 crc kubenswrapper[4824]: I0121 11:27:51.943596 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:51 crc kubenswrapper[4824]: I0121 11:27:51.963429 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.009054 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.248714 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.344490 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle\") pod \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.344561 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rs4n\" (UniqueName: \"kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n\") pod \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.344669 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data\") pod \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.344710 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts\") pod \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.350566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts" (OuterVolumeSpecName: "scripts") pod "9c73eda8-d376-4e18-9b33-f5ac4d10d091" (UID: "9c73eda8-d376-4e18-9b33-f5ac4d10d091"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.358809 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n" (OuterVolumeSpecName: "kube-api-access-5rs4n") pod "9c73eda8-d376-4e18-9b33-f5ac4d10d091" (UID: "9c73eda8-d376-4e18-9b33-f5ac4d10d091"). InnerVolumeSpecName "kube-api-access-5rs4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.374555 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c73eda8-d376-4e18-9b33-f5ac4d10d091" (UID: "9c73eda8-d376-4e18-9b33-f5ac4d10d091"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.379511 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data" (OuterVolumeSpecName: "config-data") pod "9c73eda8-d376-4e18-9b33-f5ac4d10d091" (UID: "9c73eda8-d376-4e18-9b33-f5ac4d10d091"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.446380 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.446409 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.446427 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.446437 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rs4n\" (UniqueName: \"kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.510556 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.515748 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649629 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsb2r\" (UniqueName: \"kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r\") pod \"305dccfd-067c-4062-b9d1-0200e105e8fc\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649665 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom\") pod \"305dccfd-067c-4062-b9d1-0200e105e8fc\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649685 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4787\" (UniqueName: \"kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787\") pod \"f5046792-741c-487c-adb8-43502e9fdba1\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649804 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle\") pod \"305dccfd-067c-4062-b9d1-0200e105e8fc\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649832 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data\") pod \"305dccfd-067c-4062-b9d1-0200e105e8fc\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649899 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle\") pod \"f5046792-741c-487c-adb8-43502e9fdba1\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649948 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom\") pod \"f5046792-741c-487c-adb8-43502e9fdba1\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.650000 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data\") pod \"f5046792-741c-487c-adb8-43502e9fdba1\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.670092 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "305dccfd-067c-4062-b9d1-0200e105e8fc" (UID: "305dccfd-067c-4062-b9d1-0200e105e8fc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.675079 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r" (OuterVolumeSpecName: "kube-api-access-gsb2r") pod "305dccfd-067c-4062-b9d1-0200e105e8fc" (UID: "305dccfd-067c-4062-b9d1-0200e105e8fc"). InnerVolumeSpecName "kube-api-access-gsb2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.675103 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f5046792-741c-487c-adb8-43502e9fdba1" (UID: "f5046792-741c-487c-adb8-43502e9fdba1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.685821 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787" (OuterVolumeSpecName: "kube-api-access-x4787") pod "f5046792-741c-487c-adb8-43502e9fdba1" (UID: "f5046792-741c-487c-adb8-43502e9fdba1"). InnerVolumeSpecName "kube-api-access-x4787". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.758207 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5046792-741c-487c-adb8-43502e9fdba1" (UID: "f5046792-741c-487c-adb8-43502e9fdba1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.759396 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.759416 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.759426 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsb2r\" (UniqueName: \"kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.759435 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.759444 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4787\" (UniqueName: \"kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.779107 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "305dccfd-067c-4062-b9d1-0200e105e8fc" (UID: "305dccfd-067c-4062-b9d1-0200e105e8fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.779177 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.785569 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data" (OuterVolumeSpecName: "config-data") pod "f5046792-741c-487c-adb8-43502e9fdba1" (UID: "f5046792-741c-487c-adb8-43502e9fdba1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.806292 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data" (OuterVolumeSpecName: "config-data") pod "305dccfd-067c-4062-b9d1-0200e105e8fc" (UID: "305dccfd-067c-4062-b9d1-0200e105e8fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.834104 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.861683 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.861709 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.861719 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.877664 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" event={"ID":"305dccfd-067c-4062-b9d1-0200e105e8fc","Type":"ContainerDied","Data":"88223e1727fde2b87a86e2d66dfde6e4b257cc1ef3f6ba548c34ddd1445f80ea"} Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.878489 4824 scope.go:117] "RemoveContainer" containerID="215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.878616 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.882730 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.884026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b744b9d97-zkcv4" event={"ID":"f5046792-741c-487c-adb8-43502e9fdba1","Type":"ContainerDied","Data":"8b94a9402348e00a388234d7d2563fa253b97825d6e2edc2c72d291efaccf943"} Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.907248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" event={"ID":"9c73eda8-d376-4e18-9b33-f5ac4d10d091","Type":"ContainerDied","Data":"bec80828734b592cc51fee125cf162b17f1a9431a491c60a24c542b64ae9e046"} Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.907280 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bec80828734b592cc51fee125cf162b17f1a9431a491c60a24c542b64ae9e046" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.907289 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.912349 4824 scope.go:117] "RemoveContainer" containerID="e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.913367 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerDied","Data":"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc"} Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.914753 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.922126 4824 generic.go:334] "Generic (PLEG): container finished" podID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerID="1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc" exitCode=0 Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.922193 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerDied","Data":"cd894bcfd6b429a651ad2e896089b4e35850d9690bc597b40bbe1fa904affe57"} Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.938106 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.949650 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.951867 4824 scope.go:117] "RemoveContainer" containerID="526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.963588 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-928nq\" (UniqueName: \"kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.963685 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.963756 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.963810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.963927 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.964019 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.964086 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.964262 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.964314 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.964540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.973121 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.973142 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.975103 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts" (OuterVolumeSpecName: "scripts") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.977310 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.979168 4824 scope.go:117] "RemoveContainer" containerID="046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.985846 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq" (OuterVolumeSpecName: "kube-api-access-928nq") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "kube-api-access-928nq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.988915 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989500 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="sg-core" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989513 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="sg-core" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989525 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-notification-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989532 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-notification-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989542 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="dnsmasq-dns" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989547 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="dnsmasq-dns" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989554 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="init" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989559 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="init" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989573 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989578 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989585 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989590 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989597 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="proxy-httpd" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989602 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="proxy-httpd" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989611 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989616 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989626 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c73eda8-d376-4e18-9b33-f5ac4d10d091" containerName="nova-cell0-conductor-db-sync" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989631 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c73eda8-d376-4e18-9b33-f5ac4d10d091" containerName="nova-cell0-conductor-db-sync" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989638 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989643 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989653 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-central-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989659 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-central-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989673 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989678 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989849 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989859 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="proxy-httpd" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989866 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="sg-core" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989876 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989888 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c73eda8-d376-4e18-9b33-f5ac4d10d091" containerName="nova-cell0-conductor-db-sync" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989897 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989908 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989915 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="dnsmasq-dns" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989923 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989936 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-central-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989945 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-notification-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.990491 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.992393 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.992777 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gbh62" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.993419 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.001626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.003375 4824 scope.go:117] "RemoveContainer" containerID="2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.027931 4824 scope.go:117] "RemoveContainer" containerID="1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.048510 4824 scope.go:117] "RemoveContainer" containerID="526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6" Jan 21 11:27:53 crc kubenswrapper[4824]: E0121 11:27:53.049306 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6\": container with ID starting with 526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6 not found: ID does not exist" containerID="526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.049345 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6"} err="failed to get container status \"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6\": rpc error: code = NotFound desc = could not find container \"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6\": container with ID starting with 526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6 not found: ID does not exist" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.049364 4824 scope.go:117] "RemoveContainer" containerID="046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2" Jan 21 11:27:53 crc kubenswrapper[4824]: E0121 11:27:53.049592 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2\": container with ID starting with 046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2 not found: ID does not exist" containerID="046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.049613 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2"} err="failed to get container status \"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2\": rpc error: code = NotFound desc = could not find container \"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2\": container with ID starting with 046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2 not found: ID does not exist" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.049626 4824 scope.go:117] "RemoveContainer" containerID="2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494" Jan 21 11:27:53 crc kubenswrapper[4824]: E0121 11:27:53.049993 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494\": container with ID starting with 2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494 not found: ID does not exist" containerID="2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.050028 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494"} err="failed to get container status \"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494\": rpc error: code = NotFound desc = could not find container \"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494\": container with ID starting with 2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494 not found: ID does not exist" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.050051 4824 scope.go:117] "RemoveContainer" containerID="1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc" Jan 21 11:27:53 crc kubenswrapper[4824]: E0121 11:27:53.050323 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc\": container with ID starting with 1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc not found: ID does not exist" containerID="1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.050344 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc"} err="failed to get container status \"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc\": rpc error: code = NotFound desc = could not find container \"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc\": container with ID starting with 1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc not found: ID does not exist" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.052097 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data" (OuterVolumeSpecName: "config-data") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.064926 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076609 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddqrh\" (UniqueName: \"kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076733 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076893 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076909 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076918 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076926 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076934 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-928nq\" (UniqueName: \"kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.178447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddqrh\" (UniqueName: \"kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.178501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.178594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.184565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.186309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.191884 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddqrh\" (UniqueName: \"kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.258340 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.264787 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.275925 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: E0121 11:27:53.276267 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.276286 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.276475 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.277893 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.279323 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.279364 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.293148 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.309247 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.384658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.384896 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zwmx\" (UniqueName: \"kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.384940 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.384975 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.385043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.385132 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.385159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487320 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487450 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487532 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zwmx\" (UniqueName: \"kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487726 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487756 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.488499 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.491946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.492476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.492923 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.502827 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.505102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zwmx\" (UniqueName: \"kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.590833 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.702866 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: W0121 11:27:53.710068 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d989da6_218c_4d56_b121_ce2002e90616.slice/crio-41bf3501b387352adac25f51221ca7a56a480e13956b916db174150a5458d50b WatchSource:0}: Error finding container 41bf3501b387352adac25f51221ca7a56a480e13956b916db174150a5458d50b: Status 404 returned error can't find the container with id 41bf3501b387352adac25f51221ca7a56a480e13956b916db174150a5458d50b Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.932087 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5d989da6-218c-4d56-b121-ce2002e90616","Type":"ContainerStarted","Data":"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c"} Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.932301 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5d989da6-218c-4d56-b121-ce2002e90616","Type":"ContainerStarted","Data":"41bf3501b387352adac25f51221ca7a56a480e13956b916db174150a5458d50b"} Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.933138 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.947550 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.947537316 podStartE2EDuration="1.947537316s" podCreationTimestamp="2026-01-21 11:27:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:53.946057164 +0000 UTC m=+1016.239086455" watchObservedRunningTime="2026-01-21 11:27:53.947537316 +0000 UTC m=+1016.240566607" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.981137 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.057143 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" path="/var/lib/kubelet/pods/305dccfd-067c-4062-b9d1-0200e105e8fc/volumes" Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.057907 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" path="/var/lib/kubelet/pods/964436a3-7963-4a72-bcb6-b7bab7e96574/volumes" Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.058619 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5046792-741c-487c-adb8-43502e9fdba1" path="/var/lib/kubelet/pods/f5046792-741c-487c-adb8-43502e9fdba1/volumes" Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.939852 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerStarted","Data":"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887"} Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.940083 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerStarted","Data":"092b4d89a4764210f21850893c608f45eeb6897569b66ec19869b2ec6f56a849"} Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.982626 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:55 crc kubenswrapper[4824]: I0121 11:27:55.947253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerStarted","Data":"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f"} Jan 21 11:27:56 crc kubenswrapper[4824]: I0121 11:27:56.012784 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:56 crc kubenswrapper[4824]: I0121 11:27:56.955338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerStarted","Data":"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6"} Jan 21 11:27:56 crc kubenswrapper[4824]: I0121 11:27:56.955434 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="5d989da6-218c-4d56-b121-ce2002e90616" containerName="nova-cell0-conductor-conductor" containerID="cri-o://30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c" gracePeriod=30 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.646596 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.753457 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddqrh\" (UniqueName: \"kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh\") pod \"5d989da6-218c-4d56-b121-ce2002e90616\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.753718 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle\") pod \"5d989da6-218c-4d56-b121-ce2002e90616\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.753827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data\") pod \"5d989da6-218c-4d56-b121-ce2002e90616\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.757218 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh" (OuterVolumeSpecName: "kube-api-access-ddqrh") pod "5d989da6-218c-4d56-b121-ce2002e90616" (UID: "5d989da6-218c-4d56-b121-ce2002e90616"). InnerVolumeSpecName "kube-api-access-ddqrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.773081 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d989da6-218c-4d56-b121-ce2002e90616" (UID: "5d989da6-218c-4d56-b121-ce2002e90616"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.774316 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data" (OuterVolumeSpecName: "config-data") pod "5d989da6-218c-4d56-b121-ce2002e90616" (UID: "5d989da6-218c-4d56-b121-ce2002e90616"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.856086 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddqrh\" (UniqueName: \"kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.856111 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.856120 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerStarted","Data":"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520"} Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964452 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-central-agent" containerID="cri-o://dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887" gracePeriod=30 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964513 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964536 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="proxy-httpd" containerID="cri-o://a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520" gracePeriod=30 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964572 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="sg-core" containerID="cri-o://c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6" gracePeriod=30 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964606 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-notification-agent" containerID="cri-o://9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f" gracePeriod=30 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.971074 4824 generic.go:334] "Generic (PLEG): container finished" podID="5d989da6-218c-4d56-b121-ce2002e90616" containerID="30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c" exitCode=0 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.971103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5d989da6-218c-4d56-b121-ce2002e90616","Type":"ContainerDied","Data":"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c"} Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.971128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5d989da6-218c-4d56-b121-ce2002e90616","Type":"ContainerDied","Data":"41bf3501b387352adac25f51221ca7a56a480e13956b916db174150a5458d50b"} Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.971140 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.971153 4824 scope.go:117] "RemoveContainer" containerID="30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.989736 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.524160698 podStartE2EDuration="4.989723109s" podCreationTimestamp="2026-01-21 11:27:53 +0000 UTC" firstStartedPulling="2026-01-21 11:27:53.983759258 +0000 UTC m=+1016.276788551" lastFinishedPulling="2026-01-21 11:27:57.44932168 +0000 UTC m=+1019.742350962" observedRunningTime="2026-01-21 11:27:57.986120375 +0000 UTC m=+1020.279149667" watchObservedRunningTime="2026-01-21 11:27:57.989723109 +0000 UTC m=+1020.282752402" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.000094 4824 scope.go:117] "RemoveContainer" containerID="30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c" Jan 21 11:27:58 crc kubenswrapper[4824]: E0121 11:27:58.000810 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c\": container with ID starting with 30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c not found: ID does not exist" containerID="30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.000849 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c"} err="failed to get container status \"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c\": rpc error: code = NotFound desc = could not find container \"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c\": container with ID starting with 30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c not found: ID does not exist" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.007163 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.021547 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.047591 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:58 crc kubenswrapper[4824]: E0121 11:27:58.050677 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d989da6-218c-4d56-b121-ce2002e90616" containerName="nova-cell0-conductor-conductor" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.050706 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d989da6-218c-4d56-b121-ce2002e90616" containerName="nova-cell0-conductor-conductor" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.051326 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d989da6-218c-4d56-b121-ce2002e90616" containerName="nova-cell0-conductor-conductor" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.052377 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.060663 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gbh62" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.063318 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.064186 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d989da6-218c-4d56-b121-ce2002e90616" path="/var/lib/kubelet/pods/5d989da6-218c-4d56-b121-ce2002e90616/volumes" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.064641 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.160567 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x989p\" (UniqueName: \"kubernetes.io/projected/8effdacf-8fec-40b4-8f61-8856ac0f232c-kube-api-access-x989p\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.160882 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.161073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.262636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.262758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.262927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x989p\" (UniqueName: \"kubernetes.io/projected/8effdacf-8fec-40b4-8f61-8856ac0f232c-kube-api-access-x989p\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.266374 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.273813 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.275765 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x989p\" (UniqueName: \"kubernetes.io/projected/8effdacf-8fec-40b4-8f61-8856ac0f232c-kube-api-access-x989p\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.374319 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.748440 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:58 crc kubenswrapper[4824]: W0121 11:27:58.754877 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8effdacf_8fec_40b4_8f61_8856ac0f232c.slice/crio-25998cf7bd255205c22a0312ba7cdf9b107c0a619888da32f9b0e42fd7815107 WatchSource:0}: Error finding container 25998cf7bd255205c22a0312ba7cdf9b107c0a619888da32f9b0e42fd7815107: Status 404 returned error can't find the container with id 25998cf7bd255205c22a0312ba7cdf9b107c0a619888da32f9b0e42fd7815107 Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.980190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8effdacf-8fec-40b4-8f61-8856ac0f232c","Type":"ContainerStarted","Data":"cb4175471f7b293199c781fd7c6f75c7749c4e9253027bc4a10e5ee9ff6b8076"} Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.980230 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8effdacf-8fec-40b4-8f61-8856ac0f232c","Type":"ContainerStarted","Data":"25998cf7bd255205c22a0312ba7cdf9b107c0a619888da32f9b0e42fd7815107"} Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.980415 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982690 4824 generic.go:334] "Generic (PLEG): container finished" podID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerID="a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520" exitCode=0 Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982716 4824 generic.go:334] "Generic (PLEG): container finished" podID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerID="c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6" exitCode=2 Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982726 4824 generic.go:334] "Generic (PLEG): container finished" podID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerID="9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f" exitCode=0 Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982753 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerDied","Data":"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520"} Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982784 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerDied","Data":"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6"} Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982795 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerDied","Data":"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f"} Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.998545 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=0.998532424 podStartE2EDuration="998.532424ms" podCreationTimestamp="2026-01-21 11:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:58.99225283 +0000 UTC m=+1021.285282122" watchObservedRunningTime="2026-01-21 11:27:58.998532424 +0000 UTC m=+1021.291561717" Jan 21 11:27:59 crc kubenswrapper[4824]: I0121 11:27:59.718680 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:59 crc kubenswrapper[4824]: I0121 11:27:59.756717 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:27:59 crc kubenswrapper[4824]: I0121 11:27:59.756907 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-656b845fb4-qst2s" podUID="d8d66537-20e1-4f87-9077-484e264bae81" containerName="heat-engine" containerID="cri-o://05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" gracePeriod=60 Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.720754 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823226 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823275 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823332 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823386 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823450 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823478 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zwmx\" (UniqueName: \"kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823497 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823685 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.824165 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.824181 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.832884 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx" (OuterVolumeSpecName: "kube-api-access-6zwmx") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "kube-api-access-6zwmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.838102 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts" (OuterVolumeSpecName: "scripts") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.855102 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.891639 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.905301 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data" (OuterVolumeSpecName: "config-data") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.925987 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.926011 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.926021 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.926029 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zwmx\" (UniqueName: \"kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.926040 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.004632 4824 generic.go:334] "Generic (PLEG): container finished" podID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerID="dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887" exitCode=0 Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.004669 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerDied","Data":"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887"} Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.004693 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerDied","Data":"092b4d89a4764210f21850893c608f45eeb6897569b66ec19869b2ec6f56a849"} Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.004707 4824 scope.go:117] "RemoveContainer" containerID="a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.004818 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.034678 4824 scope.go:117] "RemoveContainer" containerID="c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.041842 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.052704 4824 scope.go:117] "RemoveContainer" containerID="9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.065678 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.066739 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.067062 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-notification-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067080 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-notification-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.067094 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-central-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067100 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-central-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.067108 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="sg-core" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067113 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="sg-core" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.067132 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="proxy-httpd" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067138 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="proxy-httpd" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067309 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="sg-core" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067319 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-central-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067332 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-notification-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067347 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="proxy-httpd" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.068715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.073260 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.073503 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.074204 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.081030 4824 scope.go:117] "RemoveContainer" containerID="dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.101460 4824 scope.go:117] "RemoveContainer" containerID="a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.101748 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520\": container with ID starting with a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520 not found: ID does not exist" containerID="a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.101812 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520"} err="failed to get container status \"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520\": rpc error: code = NotFound desc = could not find container \"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520\": container with ID starting with a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520 not found: ID does not exist" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.101832 4824 scope.go:117] "RemoveContainer" containerID="c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.102066 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6\": container with ID starting with c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6 not found: ID does not exist" containerID="c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.102091 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6"} err="failed to get container status \"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6\": rpc error: code = NotFound desc = could not find container \"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6\": container with ID starting with c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6 not found: ID does not exist" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.102105 4824 scope.go:117] "RemoveContainer" containerID="9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.102319 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f\": container with ID starting with 9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f not found: ID does not exist" containerID="9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.102339 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f"} err="failed to get container status \"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f\": rpc error: code = NotFound desc = could not find container \"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f\": container with ID starting with 9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f not found: ID does not exist" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.102353 4824 scope.go:117] "RemoveContainer" containerID="dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.102595 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887\": container with ID starting with dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887 not found: ID does not exist" containerID="dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.102613 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887"} err="failed to get container status \"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887\": rpc error: code = NotFound desc = could not find container \"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887\": container with ID starting with dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887 not found: ID does not exist" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.230463 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.230813 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.230874 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.230921 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.231054 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.231092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d25xm\" (UniqueName: \"kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.231207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.332984 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333389 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333466 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d25xm\" (UniqueName: \"kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.334128 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.334163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.336591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.336635 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.337081 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.337185 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.347383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d25xm\" (UniqueName: \"kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.395569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.687293 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.688419 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.689717 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.689750 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-656b845fb4-qst2s" podUID="d8d66537-20e1-4f87-9077-484e264bae81" containerName="heat-engine" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.787486 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.014702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerStarted","Data":"bac5f30d5f4848de2bccab1913e20b18c8655df8b33e352d91fc9c00699884ca"} Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.401835 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.795841 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.967995 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-g9vq5"] Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.969615 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.977550 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.977796 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.990424 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9vq5"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.030860 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerStarted","Data":"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e"} Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.068754 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" path="/var/lib/kubelet/pods/4229a37f-5aa8-46c7-8ec4-74dde8937047/volumes" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.077295 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.077375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.077580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.077620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xnkr\" (UniqueName: \"kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.081890 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.103004 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.103088 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.104315 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.105407 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.109597 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.109832 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.159433 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.178124 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.179262 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.185004 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.191697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.192384 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xnkr\" (UniqueName: \"kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.192549 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.192677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.196745 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.197935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.216427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.231327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xnkr\" (UniqueName: \"kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.289920 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.294130 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295288 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znk42\" (UniqueName: \"kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295325 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295348 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295381 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjvlf\" (UniqueName: \"kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295438 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295462 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9drq\" (UniqueName: \"kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295550 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.337576 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.338696 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.341809 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.357885 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.373754 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.375233 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.392531 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397097 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397143 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znk42\" (UniqueName: \"kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397344 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjvlf\" (UniqueName: \"kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397458 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397500 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9drq\" (UniqueName: \"kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397688 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.399601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.400137 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.405771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.411268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.415224 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.415486 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znk42\" (UniqueName: \"kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.416261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.416507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.416646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjvlf\" (UniqueName: \"kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.421566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.433396 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9drq\" (UniqueName: \"kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.501873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502257 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502304 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502331 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502386 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kz5l\" (UniqueName: \"kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502488 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6l4s\" (UniqueName: \"kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502561 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.526332 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.537143 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.575281 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.605941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613176 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.606688 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613832 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613899 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kz5l\" (UniqueName: \"kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.614037 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6l4s\" (UniqueName: \"kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.614121 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.614166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.614480 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.614720 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.615319 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.618517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.620297 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.632372 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kz5l\" (UniqueName: \"kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.635261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6l4s\" (UniqueName: \"kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.674479 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.825723 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.851443 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9vq5"] Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.067527 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9vq5" event={"ID":"baa34da9-3711-4cd4-a5eb-ee5814ac51a7","Type":"ContainerStarted","Data":"b818724198e29357d179e92d306ffc141219f83ff62bb7dd6d3881b003b27078"} Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.072830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerStarted","Data":"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6"} Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.122350 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.150942 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.271809 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.436914 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.603035 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.114465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerStarted","Data":"2c15246588a81cb1f14c8e607d836b282630b2c662037ff7639bacf65e5fa2ae"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.116807 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerStarted","Data":"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.120651 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6","Type":"ContainerStarted","Data":"720379547d053b3767f55293ad70e28fa03d8b901bf9feb0905f084febf1fbbe"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.126220 4824 generic.go:334] "Generic (PLEG): container finished" podID="8247185f-987d-4f6e-b633-042907078e39" containerID="3f015c1fd2be4ed1835afc4169552ab51d97958ba2fb966dc794c0c63861420b" exitCode=0 Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.126267 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" event={"ID":"8247185f-987d-4f6e-b633-042907078e39","Type":"ContainerDied","Data":"3f015c1fd2be4ed1835afc4169552ab51d97958ba2fb966dc794c0c63861420b"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.126286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" event={"ID":"8247185f-987d-4f6e-b633-042907078e39","Type":"ContainerStarted","Data":"cee4d42d6e8f113c3a920c0a14aa2caa944d9803b04e59596b4d760c6b1d3e91"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.136173 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerStarted","Data":"d793c49926c6cf8bd7a478c6e30627a0b61508eda4904423a7dd0b8012e4b62b"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.138190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d373f159-2b64-4dd6-9169-a96a01afcfce","Type":"ContainerStarted","Data":"7669fbc7cdadc946c7f940b38d316891febbd77d056792c26182131fc73631b1"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.139866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9vq5" event={"ID":"baa34da9-3711-4cd4-a5eb-ee5814ac51a7","Type":"ContainerStarted","Data":"dd76c88e84c244299a1fbb85cad5296bd9ab946000d0d268303768f60e22c9c4"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.525664 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-g9vq5" podStartSLOduration=3.5256496950000002 podStartE2EDuration="3.525649695s" podCreationTimestamp="2026-01-21 11:28:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:06.170845113 +0000 UTC m=+1028.463874406" watchObservedRunningTime="2026-01-21 11:28:06.525649695 +0000 UTC m=+1028.818678988" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.576378 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7gtcg"] Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.577487 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.581551 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.588372 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.597850 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7gtcg"] Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.683123 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv6bv\" (UniqueName: \"kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.683222 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.683246 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.683309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.785888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.785939 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.786020 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.786147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv6bv\" (UniqueName: \"kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.802880 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.808550 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.810403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv6bv\" (UniqueName: \"kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.811485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.948039 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157341 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerStarted","Data":"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82"} Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157513 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-central-agent" containerID="cri-o://b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e" gracePeriod=30 Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157776 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="proxy-httpd" containerID="cri-o://70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82" gracePeriod=30 Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157811 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157836 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="sg-core" containerID="cri-o://64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3" gracePeriod=30 Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157890 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-notification-agent" containerID="cri-o://03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6" gracePeriod=30 Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.170565 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" event={"ID":"8247185f-987d-4f6e-b633-042907078e39","Type":"ContainerStarted","Data":"32fb68e8168ce229b5eee2857a8b6f19fa0565f1c232d572bde25ae83e70fb37"} Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.179905 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.638577278 podStartE2EDuration="5.17989422s" podCreationTimestamp="2026-01-21 11:28:02 +0000 UTC" firstStartedPulling="2026-01-21 11:28:02.790558451 +0000 UTC m=+1025.083587743" lastFinishedPulling="2026-01-21 11:28:06.331875393 +0000 UTC m=+1028.624904685" observedRunningTime="2026-01-21 11:28:07.17479897 +0000 UTC m=+1029.467828262" watchObservedRunningTime="2026-01-21 11:28:07.17989422 +0000 UTC m=+1029.472923512" Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.182624 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.517771 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" podStartSLOduration=3.517756134 podStartE2EDuration="3.517756134s" podCreationTimestamp="2026-01-21 11:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:07.206919767 +0000 UTC m=+1029.499949059" watchObservedRunningTime="2026-01-21 11:28:07.517756134 +0000 UTC m=+1029.810785427" Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.523437 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.530024 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7gtcg"] Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.536333 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.182209 4824 generic.go:334] "Generic (PLEG): container finished" podID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerID="70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82" exitCode=0 Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.182401 4824 generic.go:334] "Generic (PLEG): container finished" podID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerID="64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3" exitCode=2 Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.182410 4824 generic.go:334] "Generic (PLEG): container finished" podID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerID="03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6" exitCode=0 Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.183255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerDied","Data":"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82"} Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.183279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerDied","Data":"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3"} Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.183290 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerDied","Data":"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6"} Jan 21 11:28:08 crc kubenswrapper[4824]: W0121 11:28:08.708314 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod135a8573_b55a_4c5f_9cb2_a7c3adea9720.slice/crio-74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df WatchSource:0}: Error finding container 74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df: Status 404 returned error can't find the container with id 74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df Jan 21 11:28:09 crc kubenswrapper[4824]: I0121 11:28:09.193759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" event={"ID":"135a8573-b55a-4c5f-9cb2-a7c3adea9720","Type":"ContainerStarted","Data":"74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.202652 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerStarted","Data":"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.202883 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerStarted","Data":"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.202770 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-metadata" containerID="cri-o://3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" gracePeriod=30 Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.202711 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-log" containerID="cri-o://1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" gracePeriod=30 Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.204840 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d373f159-2b64-4dd6-9169-a96a01afcfce","Type":"ContainerStarted","Data":"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.204918 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="d373f159-2b64-4dd6-9169-a96a01afcfce" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f" gracePeriod=30 Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.209630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerStarted","Data":"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.209666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerStarted","Data":"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.211204 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" event={"ID":"135a8573-b55a-4c5f-9cb2-a7c3adea9720","Type":"ContainerStarted","Data":"e6d7bd7f99afa7be3d7eee7ba3147eedee78587b76d733c8cb546689d1b785fe"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.215292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6","Type":"ContainerStarted","Data":"e871bc06d477443b88ec205254bfc60975ceca19ed6da8e295be1de5bf3e986e"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.224628 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.233256646 podStartE2EDuration="6.224613476s" podCreationTimestamp="2026-01-21 11:28:04 +0000 UTC" firstStartedPulling="2026-01-21 11:28:05.161974974 +0000 UTC m=+1027.455004266" lastFinishedPulling="2026-01-21 11:28:09.153331803 +0000 UTC m=+1031.446361096" observedRunningTime="2026-01-21 11:28:10.219770541 +0000 UTC m=+1032.512799832" watchObservedRunningTime="2026-01-21 11:28:10.224613476 +0000 UTC m=+1032.517642767" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.234780 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.544110698 podStartE2EDuration="6.234763501s" podCreationTimestamp="2026-01-21 11:28:04 +0000 UTC" firstStartedPulling="2026-01-21 11:28:05.45758397 +0000 UTC m=+1027.750613262" lastFinishedPulling="2026-01-21 11:28:09.148236773 +0000 UTC m=+1031.441266065" observedRunningTime="2026-01-21 11:28:10.233325138 +0000 UTC m=+1032.526354430" watchObservedRunningTime="2026-01-21 11:28:10.234763501 +0000 UTC m=+1032.527792792" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.251802 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.264828487 podStartE2EDuration="6.25177563s" podCreationTimestamp="2026-01-21 11:28:04 +0000 UTC" firstStartedPulling="2026-01-21 11:28:05.163597966 +0000 UTC m=+1027.456627257" lastFinishedPulling="2026-01-21 11:28:09.150545108 +0000 UTC m=+1031.443574400" observedRunningTime="2026-01-21 11:28:10.247140707 +0000 UTC m=+1032.540169999" watchObservedRunningTime="2026-01-21 11:28:10.25177563 +0000 UTC m=+1032.544804921" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.260262 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" podStartSLOduration=4.260245456 podStartE2EDuration="4.260245456s" podCreationTimestamp="2026-01-21 11:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:10.257749247 +0000 UTC m=+1032.550778539" watchObservedRunningTime="2026-01-21 11:28:10.260245456 +0000 UTC m=+1032.553274748" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.286980 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.421375541 podStartE2EDuration="6.28695164s" podCreationTimestamp="2026-01-21 11:28:04 +0000 UTC" firstStartedPulling="2026-01-21 11:28:05.284995309 +0000 UTC m=+1027.578024600" lastFinishedPulling="2026-01-21 11:28:09.150571407 +0000 UTC m=+1031.443600699" observedRunningTime="2026-01-21 11:28:10.269593748 +0000 UTC m=+1032.562623040" watchObservedRunningTime="2026-01-21 11:28:10.28695164 +0000 UTC m=+1032.579980932" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.710151 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.773689 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle\") pod \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.773767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs\") pod \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.773833 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data\") pod \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.773909 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9drq\" (UniqueName: \"kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq\") pod \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.774381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs" (OuterVolumeSpecName: "logs") pod "f3a7f750-590d-43f3-a243-ec2548d1d4d8" (UID: "f3a7f750-590d-43f3-a243-ec2548d1d4d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.790121 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq" (OuterVolumeSpecName: "kube-api-access-l9drq") pod "f3a7f750-590d-43f3-a243-ec2548d1d4d8" (UID: "f3a7f750-590d-43f3-a243-ec2548d1d4d8"). InnerVolumeSpecName "kube-api-access-l9drq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.799327 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3a7f750-590d-43f3-a243-ec2548d1d4d8" (UID: "f3a7f750-590d-43f3-a243-ec2548d1d4d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.802161 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data" (OuterVolumeSpecName: "config-data") pod "f3a7f750-590d-43f3-a243-ec2548d1d4d8" (UID: "f3a7f750-590d-43f3-a243-ec2548d1d4d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.876236 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.876282 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.876292 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9drq\" (UniqueName: \"kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.876302 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223465 4824 generic.go:334] "Generic (PLEG): container finished" podID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerID="3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" exitCode=0 Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223668 4824 generic.go:334] "Generic (PLEG): container finished" podID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerID="1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" exitCode=143 Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223536 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerDied","Data":"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0"} Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223891 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerDied","Data":"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e"} Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223915 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerDied","Data":"d793c49926c6cf8bd7a478c6e30627a0b61508eda4904423a7dd0b8012e4b62b"} Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223930 4824 scope.go:117] "RemoveContainer" containerID="3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223517 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.250516 4824 scope.go:117] "RemoveContainer" containerID="1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.256997 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.269043 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.274339 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:11 crc kubenswrapper[4824]: E0121 11:28:11.274917 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-log" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.274938 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-log" Jan 21 11:28:11 crc kubenswrapper[4824]: E0121 11:28:11.274973 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-metadata" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.274979 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-metadata" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.275345 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-metadata" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.275383 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-log" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.277324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.282411 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.282588 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.286450 4824 scope.go:117] "RemoveContainer" containerID="3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.286851 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:11 crc kubenswrapper[4824]: E0121 11:28:11.287023 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0\": container with ID starting with 3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0 not found: ID does not exist" containerID="3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.287057 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0"} err="failed to get container status \"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0\": rpc error: code = NotFound desc = could not find container \"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0\": container with ID starting with 3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0 not found: ID does not exist" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.287080 4824 scope.go:117] "RemoveContainer" containerID="1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" Jan 21 11:28:11 crc kubenswrapper[4824]: E0121 11:28:11.290274 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e\": container with ID starting with 1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e not found: ID does not exist" containerID="1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.290299 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e"} err="failed to get container status \"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e\": rpc error: code = NotFound desc = could not find container \"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e\": container with ID starting with 1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e not found: ID does not exist" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.290315 4824 scope.go:117] "RemoveContainer" containerID="3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.291531 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0"} err="failed to get container status \"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0\": rpc error: code = NotFound desc = could not find container \"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0\": container with ID starting with 3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0 not found: ID does not exist" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.291552 4824 scope.go:117] "RemoveContainer" containerID="1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.293336 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e"} err="failed to get container status \"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e\": rpc error: code = NotFound desc = could not find container \"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e\": container with ID starting with 1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e not found: ID does not exist" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.385983 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.386069 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.386092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.386117 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.386444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m24v8\" (UniqueName: \"kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m24v8\" (UniqueName: \"kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489775 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489854 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.493273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.493734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.495251 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.504746 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m24v8\" (UniqueName: \"kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.662110 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.667544 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.767021 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.794368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data\") pod \"d8d66537-20e1-4f87-9077-484e264bae81\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.794537 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle\") pod \"d8d66537-20e1-4f87-9077-484e264bae81\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.794594 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom\") pod \"d8d66537-20e1-4f87-9077-484e264bae81\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.794623 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhf6r\" (UniqueName: \"kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r\") pod \"d8d66537-20e1-4f87-9077-484e264bae81\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.798401 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r" (OuterVolumeSpecName: "kube-api-access-rhf6r") pod "d8d66537-20e1-4f87-9077-484e264bae81" (UID: "d8d66537-20e1-4f87-9077-484e264bae81"). InnerVolumeSpecName "kube-api-access-rhf6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.799393 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhf6r\" (UniqueName: \"kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.807942 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d8d66537-20e1-4f87-9077-484e264bae81" (UID: "d8d66537-20e1-4f87-9077-484e264bae81"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.826596 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8d66537-20e1-4f87-9077-484e264bae81" (UID: "d8d66537-20e1-4f87-9077-484e264bae81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.840596 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data" (OuterVolumeSpecName: "config-data") pod "d8d66537-20e1-4f87-9077-484e264bae81" (UID: "d8d66537-20e1-4f87-9077-484e264bae81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.900974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901361 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901421 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901445 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901357 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901597 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901643 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d25xm\" (UniqueName: \"kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901884 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.902726 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.902747 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.902758 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.902766 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.902774 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.904566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts" (OuterVolumeSpecName: "scripts") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.909472 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm" (OuterVolumeSpecName: "kube-api-access-d25xm") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "kube-api-access-d25xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.927227 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.961370 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.987144 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data" (OuterVolumeSpecName: "config-data") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.004238 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.004272 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d25xm\" (UniqueName: \"kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.004286 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.004295 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.004303 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.057873 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" path="/var/lib/kubelet/pods/f3a7f750-590d-43f3-a243-ec2548d1d4d8/volumes" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.086428 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:12 crc kubenswrapper[4824]: W0121 11:28:12.092658 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9372ff5e_019e_47b0_b2b1_fd3f5cbd16f3.slice/crio-3b22de5e8cf949114c3b9b5d7ecfe3c1b6ca7ce93636a1e49cddaaa4c2711814 WatchSource:0}: Error finding container 3b22de5e8cf949114c3b9b5d7ecfe3c1b6ca7ce93636a1e49cddaaa4c2711814: Status 404 returned error can't find the container with id 3b22de5e8cf949114c3b9b5d7ecfe3c1b6ca7ce93636a1e49cddaaa4c2711814 Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.232531 4824 generic.go:334] "Generic (PLEG): container finished" podID="baa34da9-3711-4cd4-a5eb-ee5814ac51a7" containerID="dd76c88e84c244299a1fbb85cad5296bd9ab946000d0d268303768f60e22c9c4" exitCode=0 Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.232595 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9vq5" event={"ID":"baa34da9-3711-4cd4-a5eb-ee5814ac51a7","Type":"ContainerDied","Data":"dd76c88e84c244299a1fbb85cad5296bd9ab946000d0d268303768f60e22c9c4"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.235107 4824 generic.go:334] "Generic (PLEG): container finished" podID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerID="b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e" exitCode=0 Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.235143 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerDied","Data":"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.235179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerDied","Data":"bac5f30d5f4848de2bccab1913e20b18c8655df8b33e352d91fc9c00699884ca"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.235197 4824 scope.go:117] "RemoveContainer" containerID="70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.235203 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.239561 4824 generic.go:334] "Generic (PLEG): container finished" podID="d8d66537-20e1-4f87-9077-484e264bae81" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" exitCode=0 Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.239742 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.239762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-656b845fb4-qst2s" event={"ID":"d8d66537-20e1-4f87-9077-484e264bae81","Type":"ContainerDied","Data":"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.239833 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-656b845fb4-qst2s" event={"ID":"d8d66537-20e1-4f87-9077-484e264bae81","Type":"ContainerDied","Data":"6515b7c77086397376de7620cef207d703b735c2b9d2c10ad1793aedbc3eb3b1"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.243077 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerStarted","Data":"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.243103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerStarted","Data":"3b22de5e8cf949114c3b9b5d7ecfe3c1b6ca7ce93636a1e49cddaaa4c2711814"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.259688 4824 scope.go:117] "RemoveContainer" containerID="64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.264511 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.271270 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.281580 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.289133 4824 scope.go:117] "RemoveContainer" containerID="03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.289506 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.297919 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.298254 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-notification-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298271 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-notification-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.298288 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="proxy-httpd" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298294 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="proxy-httpd" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.298312 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-central-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298318 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-central-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.298325 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="sg-core" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298330 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="sg-core" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.298343 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d66537-20e1-4f87-9077-484e264bae81" containerName="heat-engine" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298349 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d66537-20e1-4f87-9077-484e264bae81" containerName="heat-engine" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298535 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d66537-20e1-4f87-9077-484e264bae81" containerName="heat-engine" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298550 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-notification-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298557 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="proxy-httpd" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298570 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="sg-core" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298579 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-central-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.299993 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.304941 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.305032 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.336113 4824 scope.go:117] "RemoveContainer" containerID="b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.345350 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.362700 4824 scope.go:117] "RemoveContainer" containerID="70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.363117 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82\": container with ID starting with 70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82 not found: ID does not exist" containerID="70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.363154 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82"} err="failed to get container status \"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82\": rpc error: code = NotFound desc = could not find container \"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82\": container with ID starting with 70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82 not found: ID does not exist" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.363180 4824 scope.go:117] "RemoveContainer" containerID="64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.363481 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3\": container with ID starting with 64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3 not found: ID does not exist" containerID="64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.363505 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3"} err="failed to get container status \"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3\": rpc error: code = NotFound desc = could not find container \"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3\": container with ID starting with 64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3 not found: ID does not exist" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.363521 4824 scope.go:117] "RemoveContainer" containerID="03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.364852 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6\": container with ID starting with 03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6 not found: ID does not exist" containerID="03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.364888 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6"} err="failed to get container status \"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6\": rpc error: code = NotFound desc = could not find container \"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6\": container with ID starting with 03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6 not found: ID does not exist" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.364909 4824 scope.go:117] "RemoveContainer" containerID="b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.365175 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e\": container with ID starting with b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e not found: ID does not exist" containerID="b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.365202 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e"} err="failed to get container status \"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e\": rpc error: code = NotFound desc = could not find container \"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e\": container with ID starting with b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e not found: ID does not exist" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.365217 4824 scope.go:117] "RemoveContainer" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.382246 4824 scope.go:117] "RemoveContainer" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.382534 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea\": container with ID starting with 05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea not found: ID does not exist" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.382564 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea"} err="failed to get container status \"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea\": rpc error: code = NotFound desc = could not find container \"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea\": container with ID starting with 05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea not found: ID does not exist" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.419861 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.419902 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.419988 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.420026 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.420063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmqz4\" (UniqueName: \"kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.420261 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.420310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.522338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.522599 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.522725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmqz4\" (UniqueName: \"kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.522937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.523032 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.523035 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.523314 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.523403 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.523692 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.526469 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.526580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.526834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.526993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.538617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmqz4\" (UniqueName: \"kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.628777 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.995172 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.253906 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerStarted","Data":"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff"} Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.256585 4824 generic.go:334] "Generic (PLEG): container finished" podID="135a8573-b55a-4c5f-9cb2-a7c3adea9720" containerID="e6d7bd7f99afa7be3d7eee7ba3147eedee78587b76d733c8cb546689d1b785fe" exitCode=0 Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.256658 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" event={"ID":"135a8573-b55a-4c5f-9cb2-a7c3adea9720","Type":"ContainerDied","Data":"e6d7bd7f99afa7be3d7eee7ba3147eedee78587b76d733c8cb546689d1b785fe"} Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.259264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerStarted","Data":"3debb3f0581abba1b7346533099c933cc5a8bb94ead0c540a6932f4be37b2b39"} Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.292827 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.292813951 podStartE2EDuration="2.292813951s" podCreationTimestamp="2026-01-21 11:28:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:13.288741209 +0000 UTC m=+1035.581770502" watchObservedRunningTime="2026-01-21 11:28:13.292813951 +0000 UTC m=+1035.585843244" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.698331 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.744249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data\") pod \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.744286 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xnkr\" (UniqueName: \"kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr\") pod \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.744340 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts\") pod \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.744444 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle\") pod \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.747373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts" (OuterVolumeSpecName: "scripts") pod "baa34da9-3711-4cd4-a5eb-ee5814ac51a7" (UID: "baa34da9-3711-4cd4-a5eb-ee5814ac51a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.747639 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr" (OuterVolumeSpecName: "kube-api-access-2xnkr") pod "baa34da9-3711-4cd4-a5eb-ee5814ac51a7" (UID: "baa34da9-3711-4cd4-a5eb-ee5814ac51a7"). InnerVolumeSpecName "kube-api-access-2xnkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.763997 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data" (OuterVolumeSpecName: "config-data") pod "baa34da9-3711-4cd4-a5eb-ee5814ac51a7" (UID: "baa34da9-3711-4cd4-a5eb-ee5814ac51a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.765193 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "baa34da9-3711-4cd4-a5eb-ee5814ac51a7" (UID: "baa34da9-3711-4cd4-a5eb-ee5814ac51a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.846611 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.846849 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xnkr\" (UniqueName: \"kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.846861 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.846871 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.057426 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d66537-20e1-4f87-9077-484e264bae81" path="/var/lib/kubelet/pods/d8d66537-20e1-4f87-9077-484e264bae81/volumes" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.057992 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" path="/var/lib/kubelet/pods/deb46ee7-a757-433b-8595-ac21f9d3a03f/volumes" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.273436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerStarted","Data":"82b46e3d25bb50cb39acd7bbf34eb02ad735f71fcba98f606eba2559159f21c7"} Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.274745 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9vq5" event={"ID":"baa34da9-3711-4cd4-a5eb-ee5814ac51a7","Type":"ContainerDied","Data":"b818724198e29357d179e92d306ffc141219f83ff62bb7dd6d3881b003b27078"} Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.274779 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b818724198e29357d179e92d306ffc141219f83ff62bb7dd6d3881b003b27078" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.274780 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.414579 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.414826 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-api" containerID="cri-o://2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" gracePeriod=30 Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.414778 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-log" containerID="cri-o://79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" gracePeriod=30 Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.427045 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.433239 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" containerName="nova-scheduler-scheduler" containerID="cri-o://e871bc06d477443b88ec205254bfc60975ceca19ed6da8e295be1de5bf3e986e" gracePeriod=30 Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.437952 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.575913 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.576434 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.662971 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data\") pod \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.663059 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle\") pod \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.663083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv6bv\" (UniqueName: \"kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv\") pod \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.663220 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts\") pod \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.668172 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts" (OuterVolumeSpecName: "scripts") pod "135a8573-b55a-4c5f-9cb2-a7c3adea9720" (UID: "135a8573-b55a-4c5f-9cb2-a7c3adea9720"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.669124 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv" (OuterVolumeSpecName: "kube-api-access-vv6bv") pod "135a8573-b55a-4c5f-9cb2-a7c3adea9720" (UID: "135a8573-b55a-4c5f-9cb2-a7c3adea9720"). InnerVolumeSpecName "kube-api-access-vv6bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.675117 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.687724 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "135a8573-b55a-4c5f-9cb2-a7c3adea9720" (UID: "135a8573-b55a-4c5f-9cb2-a7c3adea9720"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.693026 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data" (OuterVolumeSpecName: "config-data") pod "135a8573-b55a-4c5f-9cb2-a7c3adea9720" (UID: "135a8573-b55a-4c5f-9cb2-a7c3adea9720"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.765623 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.765651 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.765661 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv6bv\" (UniqueName: \"kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.765671 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.829118 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.880321 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.880501 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="dnsmasq-dns" containerID="cri-o://18927e4fff1f2c120dbfd5a267daf57c4c84dbc311e5fd297d4153df2a6f21f4" gracePeriod=10 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:14.997581 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.070523 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle\") pod \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.070637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs\") pod \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.070659 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data\") pod \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.070767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjvlf\" (UniqueName: \"kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf\") pod \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.071319 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs" (OuterVolumeSpecName: "logs") pod "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" (UID: "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.078097 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf" (OuterVolumeSpecName: "kube-api-access-pjvlf") pod "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" (UID: "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8"). InnerVolumeSpecName "kube-api-access-pjvlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.111075 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data" (OuterVolumeSpecName: "config-data") pod "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" (UID: "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.127012 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" (UID: "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.174300 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.174324 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.174335 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjvlf\" (UniqueName: \"kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.174352 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.291223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" event={"ID":"135a8573-b55a-4c5f-9cb2-a7c3adea9720","Type":"ContainerDied","Data":"74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.291261 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.291279 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.293578 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerStarted","Data":"87edd33f33351d88d9cbd6b17c925a8893b337952321dc91c1ba77e5c174ae3c"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.293623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerStarted","Data":"97c437ef91be90c480bbbc66844b2f1bf63fc92b3d5efc241a8ff7639e0f322c"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.297110 4824 generic.go:334] "Generic (PLEG): container finished" podID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerID="18927e4fff1f2c120dbfd5a267daf57c4c84dbc311e5fd297d4153df2a6f21f4" exitCode=0 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.297137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" event={"ID":"f2ec0128-5881-4c67-8d78-d53d8be3a2d3","Type":"ContainerDied","Data":"18927e4fff1f2c120dbfd5a267daf57c4c84dbc311e5fd297d4153df2a6f21f4"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.299188 4824 generic.go:334] "Generic (PLEG): container finished" podID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerID="2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" exitCode=0 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.299217 4824 generic.go:334] "Generic (PLEG): container finished" podID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerID="79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" exitCode=143 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.299357 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-log" containerID="cri-o://3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" gracePeriod=30 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.299649 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.300141 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerDied","Data":"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.300189 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerDied","Data":"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.300202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerDied","Data":"2c15246588a81cb1f14c8e607d836b282630b2c662037ff7639bacf65e5fa2ae"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.300222 4824 scope.go:117] "RemoveContainer" containerID="2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.300714 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-metadata" containerID="cri-o://7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" gracePeriod=30 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.324943 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.339082 4824 scope.go:117] "RemoveContainer" containerID="79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.351097 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.363844 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374045 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374406 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baa34da9-3711-4cd4-a5eb-ee5814ac51a7" containerName="nova-manage" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374424 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="baa34da9-3711-4cd4-a5eb-ee5814ac51a7" containerName="nova-manage" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374445 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="init" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374451 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="init" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374459 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135a8573-b55a-4c5f-9cb2-a7c3adea9720" containerName="nova-cell1-conductor-db-sync" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374465 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="135a8573-b55a-4c5f-9cb2-a7c3adea9720" containerName="nova-cell1-conductor-db-sync" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374479 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-api" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374485 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-api" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374507 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="dnsmasq-dns" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374512 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="dnsmasq-dns" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374524 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-log" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374529 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-log" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374706 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-api" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374720 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="baa34da9-3711-4cd4-a5eb-ee5814ac51a7" containerName="nova-manage" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374734 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-log" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374743 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="135a8573-b55a-4c5f-9cb2-a7c3adea9720" containerName="nova-cell1-conductor-db-sync" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374752 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="dnsmasq-dns" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.375601 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380244 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380314 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380464 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380500 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380530 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380558 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hljdx\" (UniqueName: \"kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.385409 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.392137 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx" (OuterVolumeSpecName: "kube-api-access-hljdx") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "kube-api-access-hljdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.398456 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.437532 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.439192 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.441676 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config" (OuterVolumeSpecName: "config") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.442263 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.443901 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.446932 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.459312 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.463917 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.464548 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486031 4824 scope.go:117] "RemoveContainer" containerID="2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.486510 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7\": container with ID starting with 2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7 not found: ID does not exist" containerID="2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486543 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7"} err="failed to get container status \"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7\": rpc error: code = NotFound desc = could not find container \"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7\": container with ID starting with 2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7 not found: ID does not exist" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486567 4824 scope.go:117] "RemoveContainer" containerID="79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486637 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486834 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486915 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.487053 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf\": container with ID starting with 79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf not found: ID does not exist" containerID="79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487144 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5q8r\" (UniqueName: \"kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487132 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf"} err="failed to get container status \"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf\": rpc error: code = NotFound desc = could not find container \"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf\": container with ID starting with 79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf not found: ID does not exist" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487302 4824 scope.go:117] "RemoveContainer" containerID="2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487262 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487378 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487389 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487398 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487408 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487415 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hljdx\" (UniqueName: \"kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487774 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7"} err="failed to get container status \"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7\": rpc error: code = NotFound desc = could not find container \"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7\": container with ID starting with 2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7 not found: ID does not exist" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487822 4824 scope.go:117] "RemoveContainer" containerID="79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.488232 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf"} err="failed to get container status \"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf\": rpc error: code = NotFound desc = could not find container \"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf\": container with ID starting with 79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf not found: ID does not exist" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.588768 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5q8r\" (UniqueName: \"kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589047 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589504 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589588 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc5qr\" (UniqueName: \"kubernetes.io/projected/c398d1e9-13e1-400c-a609-bceea49aea51-kube-api-access-mc5qr\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.590766 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.592830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.593573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.604677 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5q8r\" (UniqueName: \"kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.691911 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.692153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.692327 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc5qr\" (UniqueName: \"kubernetes.io/projected/c398d1e9-13e1-400c-a609-bceea49aea51-kube-api-access-mc5qr\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.694920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.695017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.695335 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.712002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc5qr\" (UniqueName: \"kubernetes.io/projected/c398d1e9-13e1-400c-a609-bceea49aea51-kube-api-access-mc5qr\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.762805 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.937813 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.057506 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" path="/var/lib/kubelet/pods/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8/volumes" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.301375 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.311395 4824 generic.go:334] "Generic (PLEG): container finished" podID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" containerID="e871bc06d477443b88ec205254bfc60975ceca19ed6da8e295be1de5bf3e986e" exitCode=0 Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.311474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6","Type":"ContainerDied","Data":"e871bc06d477443b88ec205254bfc60975ceca19ed6da8e295be1de5bf3e986e"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.313083 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerStarted","Data":"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.313130 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerStarted","Data":"728845c124e990e3b5128ca1bd8dd69e1a52e3b9f1c5598a8738ea82598401a9"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320223 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerDied","Data":"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320304 4824 generic.go:334] "Generic (PLEG): container finished" podID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerID="7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" exitCode=0 Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320337 4824 scope.go:117] "RemoveContainer" containerID="7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320350 4824 generic.go:334] "Generic (PLEG): container finished" podID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerID="3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" exitCode=143 Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerDied","Data":"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320445 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerDied","Data":"3b22de5e8cf949114c3b9b5d7ecfe3c1b6ca7ce93636a1e49cddaaa4c2711814"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.325947 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" event={"ID":"f2ec0128-5881-4c67-8d78-d53d8be3a2d3","Type":"ContainerDied","Data":"76a9ebfca2b2e541f5b20e22a77a9c3eda4db32f3628a468b328bf434a816a25"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.325987 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.362932 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.373168 4824 scope.go:117] "RemoveContainer" containerID="3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.373263 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.383558 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.403278 4824 scope.go:117] "RemoveContainer" containerID="7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" Jan 21 11:28:16 crc kubenswrapper[4824]: E0121 11:28:16.404386 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff\": container with ID starting with 7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff not found: ID does not exist" containerID="7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.404437 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff"} err="failed to get container status \"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff\": rpc error: code = NotFound desc = could not find container \"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff\": container with ID starting with 7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff not found: ID does not exist" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.404458 4824 scope.go:117] "RemoveContainer" containerID="3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" Jan 21 11:28:16 crc kubenswrapper[4824]: E0121 11:28:16.406069 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540\": container with ID starting with 3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540 not found: ID does not exist" containerID="3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.406097 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540"} err="failed to get container status \"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540\": rpc error: code = NotFound desc = could not find container \"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540\": container with ID starting with 3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540 not found: ID does not exist" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.406123 4824 scope.go:117] "RemoveContainer" containerID="7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.406512 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff"} err="failed to get container status \"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff\": rpc error: code = NotFound desc = could not find container \"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff\": container with ID starting with 7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff not found: ID does not exist" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.406535 4824 scope.go:117] "RemoveContainer" containerID="3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.407387 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540"} err="failed to get container status \"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540\": rpc error: code = NotFound desc = could not find container \"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540\": container with ID starting with 3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540 not found: ID does not exist" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.407421 4824 scope.go:117] "RemoveContainer" containerID="18927e4fff1f2c120dbfd5a267daf57c4c84dbc311e5fd297d4153df2a6f21f4" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.409079 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data\") pod \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.409143 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle\") pod \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.409217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m24v8\" (UniqueName: \"kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8\") pod \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.409257 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs\") pod \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.409295 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs\") pod \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.410256 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs" (OuterVolumeSpecName: "logs") pod "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" (UID: "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.420737 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8" (OuterVolumeSpecName: "kube-api-access-m24v8") pod "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" (UID: "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3"). InnerVolumeSpecName "kube-api-access-m24v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.424570 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.444599 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" (UID: "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.453759 4824 scope.go:117] "RemoveContainer" containerID="77516fe4a72158d203b645cac966a850920a66f17eb3fb68fcdc002cb9e2b607" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.454030 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data" (OuterVolumeSpecName: "config-data") pod "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" (UID: "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.471494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" (UID: "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.510656 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data\") pod \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.510689 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znk42\" (UniqueName: \"kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42\") pod \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.510802 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle\") pod \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.511363 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.511380 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.511389 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.511398 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m24v8\" (UniqueName: \"kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.511407 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.514071 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42" (OuterVolumeSpecName: "kube-api-access-znk42") pod "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" (UID: "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6"). InnerVolumeSpecName "kube-api-access-znk42". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.547947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data" (OuterVolumeSpecName: "config-data") pod "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" (UID: "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.548007 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" (UID: "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.612609 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.612640 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.612650 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znk42\" (UniqueName: \"kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.643939 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.651548 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.689506 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: E0121 11:28:16.690177 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-metadata" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690197 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-metadata" Jan 21 11:28:16 crc kubenswrapper[4824]: E0121 11:28:16.690253 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-log" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690260 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-log" Jan 21 11:28:16 crc kubenswrapper[4824]: E0121 11:28:16.690269 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" containerName="nova-scheduler-scheduler" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690276 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" containerName="nova-scheduler-scheduler" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690516 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" containerName="nova-scheduler-scheduler" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690539 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-log" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690567 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-metadata" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.691935 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.693461 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.693842 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.695900 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.816548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.816646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.816910 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpvwd\" (UniqueName: \"kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.816952 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.817120 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.918653 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpvwd\" (UniqueName: \"kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.918708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.918779 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.918883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.918920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.919679 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.924143 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.924603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.927386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.935436 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpvwd\" (UniqueName: \"kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.010404 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.335508 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerStarted","Data":"d91972673e1ac4ea40f718ccf4ab2ec2575e9319007c3a3cc55cff5c9b8fc08a"} Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.336461 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.338407 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6","Type":"ContainerDied","Data":"720379547d053b3767f55293ad70e28fa03d8b901bf9feb0905f084febf1fbbe"} Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.338439 4824 scope.go:117] "RemoveContainer" containerID="e871bc06d477443b88ec205254bfc60975ceca19ed6da8e295be1de5bf3e986e" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.338530 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.343877 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerStarted","Data":"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254"} Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.347924 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c398d1e9-13e1-400c-a609-bceea49aea51","Type":"ContainerStarted","Data":"f51abad78825f4a03796cceaefb7a7b874eb98078ea107786fb273dcdc8422a5"} Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.348051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c398d1e9-13e1-400c-a609-bceea49aea51","Type":"ContainerStarted","Data":"4275ae8b0ade24135f7edd7238e9a5b6244c1754d3bdfca05db8d8222ab9befc"} Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.348530 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.359362 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.148877432 podStartE2EDuration="5.359350477s" podCreationTimestamp="2026-01-21 11:28:12 +0000 UTC" firstStartedPulling="2026-01-21 11:28:13.04601383 +0000 UTC m=+1035.339043122" lastFinishedPulling="2026-01-21 11:28:16.256486875 +0000 UTC m=+1038.549516167" observedRunningTime="2026-01-21 11:28:17.357813489 +0000 UTC m=+1039.650842781" watchObservedRunningTime="2026-01-21 11:28:17.359350477 +0000 UTC m=+1039.652379769" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.385580 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.385538965 podStartE2EDuration="2.385538965s" podCreationTimestamp="2026-01-21 11:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:17.381660139 +0000 UTC m=+1039.674689431" watchObservedRunningTime="2026-01-21 11:28:17.385538965 +0000 UTC m=+1039.678568257" Jan 21 11:28:17 crc kubenswrapper[4824]: W0121 11:28:17.410119 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b9162ed_f3d2_4877_b8a3_6342a25200b7.slice/crio-cf896b02cb7f72cd22b94afd5cb645eb3dc94f2424b20aa748a3b1e35979c8a8 WatchSource:0}: Error finding container cf896b02cb7f72cd22b94afd5cb645eb3dc94f2424b20aa748a3b1e35979c8a8: Status 404 returned error can't find the container with id cf896b02cb7f72cd22b94afd5cb645eb3dc94f2424b20aa748a3b1e35979c8a8 Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.413365 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.427231 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.440341 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.446838 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.447970 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.449372 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.450867 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.45085198 podStartE2EDuration="2.45085198s" podCreationTimestamp="2026-01-21 11:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:17.410184682 +0000 UTC m=+1039.703213974" watchObservedRunningTime="2026-01-21 11:28:17.45085198 +0000 UTC m=+1039.743881273" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.463698 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.528571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsphm\" (UniqueName: \"kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.528846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.529011 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.630564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.630786 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsphm\" (UniqueName: \"kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.631007 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.633565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.633809 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.644935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsphm\" (UniqueName: \"kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.763274 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.063723 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" path="/var/lib/kubelet/pods/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6/volumes" Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.064594 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" path="/var/lib/kubelet/pods/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3/volumes" Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.065152 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" path="/var/lib/kubelet/pods/f2ec0128-5881-4c67-8d78-d53d8be3a2d3/volumes" Jan 21 11:28:18 crc kubenswrapper[4824]: W0121 11:28:18.139674 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8b9cd0d_c7af_455d_a170_1521870a07f6.slice/crio-bcd965629c724abfbdcc4781034ea86d257b188520a919034221d6a8bf2c0f86 WatchSource:0}: Error finding container bcd965629c724abfbdcc4781034ea86d257b188520a919034221d6a8bf2c0f86: Status 404 returned error can't find the container with id bcd965629c724abfbdcc4781034ea86d257b188520a919034221d6a8bf2c0f86 Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.140762 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.358139 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8b9cd0d-c7af-455d-a170-1521870a07f6","Type":"ContainerStarted","Data":"be161c0756cab08e44d6825e362adf099303c412379798ba94c4f4bf705f4830"} Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.358344 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8b9cd0d-c7af-455d-a170-1521870a07f6","Type":"ContainerStarted","Data":"bcd965629c724abfbdcc4781034ea86d257b188520a919034221d6a8bf2c0f86"} Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.361691 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerStarted","Data":"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c"} Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.361728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerStarted","Data":"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9"} Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.361741 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerStarted","Data":"cf896b02cb7f72cd22b94afd5cb645eb3dc94f2424b20aa748a3b1e35979c8a8"} Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.369668 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.369655436 podStartE2EDuration="1.369655436s" podCreationTimestamp="2026-01-21 11:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:18.368504105 +0000 UTC m=+1040.661533397" watchObservedRunningTime="2026-01-21 11:28:18.369655436 +0000 UTC m=+1040.662684728" Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.384846 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.384834546 podStartE2EDuration="2.384834546s" podCreationTimestamp="2026-01-21 11:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:18.383658219 +0000 UTC m=+1040.676687511" watchObservedRunningTime="2026-01-21 11:28:18.384834546 +0000 UTC m=+1040.677863837" Jan 21 11:28:22 crc kubenswrapper[4824]: I0121 11:28:22.010934 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 21 11:28:22 crc kubenswrapper[4824]: I0121 11:28:22.011422 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 21 11:28:22 crc kubenswrapper[4824]: I0121 11:28:22.764157 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 21 11:28:25 crc kubenswrapper[4824]: I0121 11:28:25.695998 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:28:25 crc kubenswrapper[4824]: I0121 11:28:25.696346 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:28:25 crc kubenswrapper[4824]: I0121 11:28:25.783870 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:26 crc kubenswrapper[4824]: I0121 11:28:26.779093 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:26 crc kubenswrapper[4824]: I0121 11:28:26.779093 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:27 crc kubenswrapper[4824]: I0121 11:28:27.011591 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 21 11:28:27 crc kubenswrapper[4824]: I0121 11:28:27.011779 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 21 11:28:27 crc kubenswrapper[4824]: I0121 11:28:27.763620 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 21 11:28:27 crc kubenswrapper[4824]: I0121 11:28:27.788212 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 21 11:28:28 crc kubenswrapper[4824]: I0121 11:28:28.025092 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:28 crc kubenswrapper[4824]: I0121 11:28:28.025096 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:28 crc kubenswrapper[4824]: I0121 11:28:28.460270 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.698991 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.700048 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.700360 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.700390 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.703039 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.703224 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.854646 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.866500 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.886714 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.957996 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.958064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.958103 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.958484 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkm2s\" (UniqueName: \"kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.958713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.958783 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkm2s\" (UniqueName: \"kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060163 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060191 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060236 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060264 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060294 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.061065 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.061764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.062259 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.062265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.062976 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.082485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkm2s\" (UniqueName: \"kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.197998 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: W0121 11:28:36.693782 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ad8dd9e_6fe4_462b_98a9_d0192072589d.slice/crio-6738b1680c3c03cab4e1e7a4ccdc6941fbdb963c0035872cd1e033904215c51b WatchSource:0}: Error finding container 6738b1680c3c03cab4e1e7a4ccdc6941fbdb963c0035872cd1e033904215c51b: Status 404 returned error can't find the container with id 6738b1680c3c03cab4e1e7a4ccdc6941fbdb963c0035872cd1e033904215c51b Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.694451 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.017691 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.023189 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.023927 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.499029 4824 generic.go:334] "Generic (PLEG): container finished" podID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerID="ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f" exitCode=0 Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.499114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" event={"ID":"8ad8dd9e-6fe4-462b-98a9-d0192072589d","Type":"ContainerDied","Data":"ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f"} Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.499157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" event={"ID":"8ad8dd9e-6fe4-462b-98a9-d0192072589d","Type":"ContainerStarted","Data":"6738b1680c3c03cab4e1e7a4ccdc6941fbdb963c0035872cd1e033904215c51b"} Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.540617 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.686763 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.687202 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-central-agent" containerID="cri-o://82b46e3d25bb50cb39acd7bbf34eb02ad735f71fcba98f606eba2559159f21c7" gracePeriod=30 Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.687309 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="proxy-httpd" containerID="cri-o://d91972673e1ac4ea40f718ccf4ab2ec2575e9319007c3a3cc55cff5c9b8fc08a" gracePeriod=30 Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.687345 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="sg-core" containerID="cri-o://87edd33f33351d88d9cbd6b17c925a8893b337952321dc91c1ba77e5c174ae3c" gracePeriod=30 Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.687375 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-notification-agent" containerID="cri-o://97c437ef91be90c480bbbc66844b2f1bf63fc92b3d5efc241a8ff7639e0f322c" gracePeriod=30 Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.787419 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.201:3000/\": read tcp 10.217.0.2:39062->10.217.0.201:3000: read: connection reset by peer" Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.166506 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.507365 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" event={"ID":"8ad8dd9e-6fe4-462b-98a9-d0192072589d","Type":"ContainerStarted","Data":"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575"} Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.507754 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510114 4824 generic.go:334] "Generic (PLEG): container finished" podID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerID="d91972673e1ac4ea40f718ccf4ab2ec2575e9319007c3a3cc55cff5c9b8fc08a" exitCode=0 Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510143 4824 generic.go:334] "Generic (PLEG): container finished" podID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerID="87edd33f33351d88d9cbd6b17c925a8893b337952321dc91c1ba77e5c174ae3c" exitCode=2 Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510150 4824 generic.go:334] "Generic (PLEG): container finished" podID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerID="82b46e3d25bb50cb39acd7bbf34eb02ad735f71fcba98f606eba2559159f21c7" exitCode=0 Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510220 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerDied","Data":"d91972673e1ac4ea40f718ccf4ab2ec2575e9319007c3a3cc55cff5c9b8fc08a"} Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510258 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerDied","Data":"87edd33f33351d88d9cbd6b17c925a8893b337952321dc91c1ba77e5c174ae3c"} Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerDied","Data":"82b46e3d25bb50cb39acd7bbf34eb02ad735f71fcba98f606eba2559159f21c7"} Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510313 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-log" containerID="cri-o://7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249" gracePeriod=30 Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510425 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-api" containerID="cri-o://ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254" gracePeriod=30 Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.530283 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" podStartSLOduration=3.530267214 podStartE2EDuration="3.530267214s" podCreationTimestamp="2026-01-21 11:28:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:38.526274385 +0000 UTC m=+1060.819303678" watchObservedRunningTime="2026-01-21 11:28:38.530267214 +0000 UTC m=+1060.823296506" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.520087 4824 generic.go:334] "Generic (PLEG): container finished" podID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerID="97c437ef91be90c480bbbc66844b2f1bf63fc92b3d5efc241a8ff7639e0f322c" exitCode=0 Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.520330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerDied","Data":"97c437ef91be90c480bbbc66844b2f1bf63fc92b3d5efc241a8ff7639e0f322c"} Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.520353 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerDied","Data":"3debb3f0581abba1b7346533099c933cc5a8bb94ead0c540a6932f4be37b2b39"} Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.520363 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3debb3f0581abba1b7346533099c933cc5a8bb94ead0c540a6932f4be37b2b39" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.522233 4824 generic.go:334] "Generic (PLEG): container finished" podID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerID="7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249" exitCode=143 Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.522941 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerDied","Data":"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249"} Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.528935 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620459 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620520 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmqz4\" (UniqueName: \"kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620657 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620695 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620754 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620787 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620813 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.621504 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.621842 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.625355 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts" (OuterVolumeSpecName: "scripts") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.633035 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4" (OuterVolumeSpecName: "kube-api-access-dmqz4") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "kube-api-access-dmqz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.645607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.686428 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.704672 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data" (OuterVolumeSpecName: "config-data") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722638 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722664 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722675 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722683 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722691 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmqz4\" (UniqueName: \"kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722699 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722707 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.518392 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.551740 4824 generic.go:334] "Generic (PLEG): container finished" podID="d373f159-2b64-4dd6-9169-a96a01afcfce" containerID="82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f" exitCode=137 Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.551804 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.552031 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.551825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d373f159-2b64-4dd6-9169-a96a01afcfce","Type":"ContainerDied","Data":"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f"} Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.552138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d373f159-2b64-4dd6-9169-a96a01afcfce","Type":"ContainerDied","Data":"7669fbc7cdadc946c7f940b38d316891febbd77d056792c26182131fc73631b1"} Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.552164 4824 scope.go:117] "RemoveContainer" containerID="82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.577564 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.583002 4824 scope.go:117] "RemoveContainer" containerID="82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f" Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.583303 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f\": container with ID starting with 82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f not found: ID does not exist" containerID="82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.583330 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f"} err="failed to get container status \"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f\": rpc error: code = NotFound desc = could not find container \"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f\": container with ID starting with 82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f not found: ID does not exist" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.589449 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596423 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.596766 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-notification-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596778 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-notification-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.596791 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d373f159-2b64-4dd6-9169-a96a01afcfce" containerName="nova-cell1-novncproxy-novncproxy" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596809 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d373f159-2b64-4dd6-9169-a96a01afcfce" containerName="nova-cell1-novncproxy-novncproxy" Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.596838 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="proxy-httpd" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596843 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="proxy-httpd" Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.596856 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="sg-core" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596861 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="sg-core" Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.596879 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-central-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596884 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-central-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.597107 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-notification-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.597124 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="proxy-httpd" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.597130 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d373f159-2b64-4dd6-9169-a96a01afcfce" containerName="nova-cell1-novncproxy-novncproxy" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.597147 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-central-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.597155 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="sg-core" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.598970 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.600471 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.600545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.605250 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.637636 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle\") pod \"d373f159-2b64-4dd6-9169-a96a01afcfce\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.637685 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6l4s\" (UniqueName: \"kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s\") pod \"d373f159-2b64-4dd6-9169-a96a01afcfce\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.637761 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data\") pod \"d373f159-2b64-4dd6-9169-a96a01afcfce\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.648362 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s" (OuterVolumeSpecName: "kube-api-access-z6l4s") pod "d373f159-2b64-4dd6-9169-a96a01afcfce" (UID: "d373f159-2b64-4dd6-9169-a96a01afcfce"). InnerVolumeSpecName "kube-api-access-z6l4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.659138 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d373f159-2b64-4dd6-9169-a96a01afcfce" (UID: "d373f159-2b64-4dd6-9169-a96a01afcfce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.661129 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data" (OuterVolumeSpecName: "config-data") pod "d373f159-2b64-4dd6-9169-a96a01afcfce" (UID: "d373f159-2b64-4dd6-9169-a96a01afcfce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740581 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740718 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740906 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvljc\" (UniqueName: \"kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.741036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.741149 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.741165 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6l4s\" (UniqueName: \"kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.741176 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843042 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843197 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843214 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvljc\" (UniqueName: \"kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843504 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843578 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.846275 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.846464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.846524 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.846536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.856222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvljc\" (UniqueName: \"kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.876517 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.886384 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.906733 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.908573 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.915478 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.915622 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.915890 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.916086 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.922705 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.047270 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.047555 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.047639 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.047840 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2dmc\" (UniqueName: \"kubernetes.io/projected/8e42e9d6-485f-4b97-b049-316cd47af30e-kube-api-access-m2dmc\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.047966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.149265 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2dmc\" (UniqueName: \"kubernetes.io/projected/8e42e9d6-485f-4b97-b049-316cd47af30e-kube-api-access-m2dmc\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.149328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.149378 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.149450 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.149510 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.152563 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.152635 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.152852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.153758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.161925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2dmc\" (UniqueName: \"kubernetes.io/projected/8e42e9d6-485f-4b97-b049-316cd47af30e-kube-api-access-m2dmc\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.230752 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: W0121 11:28:41.318586 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55916065_772b_45ee_8871_37c229777deb.slice/crio-65e4d9ce199d66a5d4b888b55edee4667e5ba88065fc68ccd26d1ad79dbffbb0 WatchSource:0}: Error finding container 65e4d9ce199d66a5d4b888b55edee4667e5ba88065fc68ccd26d1ad79dbffbb0: Status 404 returned error can't find the container with id 65e4d9ce199d66a5d4b888b55edee4667e5ba88065fc68ccd26d1ad79dbffbb0 Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.319893 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.559513 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerStarted","Data":"65e4d9ce199d66a5d4b888b55edee4667e5ba88065fc68ccd26d1ad79dbffbb0"} Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.628614 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.948595 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.057986 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" path="/var/lib/kubelet/pods/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1/volumes" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.058818 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d373f159-2b64-4dd6-9169-a96a01afcfce" path="/var/lib/kubelet/pods/d373f159-2b64-4dd6-9169-a96a01afcfce/volumes" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.065784 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs\") pod \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.065839 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data\") pod \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.065895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle\") pod \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.065925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5q8r\" (UniqueName: \"kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r\") pod \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.066410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs" (OuterVolumeSpecName: "logs") pod "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" (UID: "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.066688 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.070046 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r" (OuterVolumeSpecName: "kube-api-access-n5q8r") pod "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" (UID: "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d"). InnerVolumeSpecName "kube-api-access-n5q8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.087096 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" (UID: "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.088110 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data" (OuterVolumeSpecName: "config-data") pod "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" (UID: "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.168193 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.168221 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5q8r\" (UniqueName: \"kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.168233 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.569164 4824 generic.go:334] "Generic (PLEG): container finished" podID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerID="ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254" exitCode=0 Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.569230 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.569255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerDied","Data":"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254"} Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.570267 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerDied","Data":"728845c124e990e3b5128ca1bd8dd69e1a52e3b9f1c5598a8738ea82598401a9"} Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.570292 4824 scope.go:117] "RemoveContainer" containerID="ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.571419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerStarted","Data":"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f"} Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.573161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e42e9d6-485f-4b97-b049-316cd47af30e","Type":"ContainerStarted","Data":"c50e2f4bbf17ec122e748d38d9b57650b3e16cc7f0d18cca04bd8713f5473e55"} Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.573191 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e42e9d6-485f-4b97-b049-316cd47af30e","Type":"ContainerStarted","Data":"faf17c84aa6350cbbfe7b1a2a465fe70e978a750b6dbf0fa539472cf64b43203"} Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.591727 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.591702482 podStartE2EDuration="2.591702482s" podCreationTimestamp="2026-01-21 11:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:42.587873542 +0000 UTC m=+1064.880902825" watchObservedRunningTime="2026-01-21 11:28:42.591702482 +0000 UTC m=+1064.884731774" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.598360 4824 scope.go:117] "RemoveContainer" containerID="7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.603968 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.621374 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.623971 4824 scope.go:117] "RemoveContainer" containerID="ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254" Jan 21 11:28:42 crc kubenswrapper[4824]: E0121 11:28:42.624792 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254\": container with ID starting with ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254 not found: ID does not exist" containerID="ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.624914 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254"} err="failed to get container status \"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254\": rpc error: code = NotFound desc = could not find container \"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254\": container with ID starting with ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254 not found: ID does not exist" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.625017 4824 scope.go:117] "RemoveContainer" containerID="7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249" Jan 21 11:28:42 crc kubenswrapper[4824]: E0121 11:28:42.625469 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249\": container with ID starting with 7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249 not found: ID does not exist" containerID="7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.625545 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249"} err="failed to get container status \"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249\": rpc error: code = NotFound desc = could not find container \"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249\": container with ID starting with 7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249 not found: ID does not exist" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.634379 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:42 crc kubenswrapper[4824]: E0121 11:28:42.634750 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-api" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.634767 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-api" Jan 21 11:28:42 crc kubenswrapper[4824]: E0121 11:28:42.634784 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-log" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.634791 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-log" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.634990 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-api" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.635013 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-log" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.635888 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.638380 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.638545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.638664 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.640196 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777384 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22h96\" (UniqueName: \"kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777454 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.878942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22h96\" (UniqueName: \"kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879093 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879189 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879540 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.882198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.882763 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.882891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.883222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.891771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22h96\" (UniqueName: \"kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.948995 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:43 crc kubenswrapper[4824]: I0121 11:28:43.348527 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:43 crc kubenswrapper[4824]: W0121 11:28:43.351291 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad597758_1f2c_4f84_8035_59fa20e68084.slice/crio-04682cea27407b0d473ba48a7b6b0c26d36ec7ff8680fb81116bbb44f989542f WatchSource:0}: Error finding container 04682cea27407b0d473ba48a7b6b0c26d36ec7ff8680fb81116bbb44f989542f: Status 404 returned error can't find the container with id 04682cea27407b0d473ba48a7b6b0c26d36ec7ff8680fb81116bbb44f989542f Jan 21 11:28:43 crc kubenswrapper[4824]: I0121 11:28:43.599398 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerStarted","Data":"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a"} Jan 21 11:28:43 crc kubenswrapper[4824]: I0121 11:28:43.599436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerStarted","Data":"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253"} Jan 21 11:28:43 crc kubenswrapper[4824]: I0121 11:28:43.601440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerStarted","Data":"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218"} Jan 21 11:28:43 crc kubenswrapper[4824]: I0121 11:28:43.601488 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerStarted","Data":"04682cea27407b0d473ba48a7b6b0c26d36ec7ff8680fb81116bbb44f989542f"} Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.056698 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" path="/var/lib/kubelet/pods/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d/volumes" Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.610272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerStarted","Data":"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74"} Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.610573 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.612343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerStarted","Data":"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331"} Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.631163 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.527669825 podStartE2EDuration="4.631148042s" podCreationTimestamp="2026-01-21 11:28:40 +0000 UTC" firstStartedPulling="2026-01-21 11:28:41.320567052 +0000 UTC m=+1063.613596344" lastFinishedPulling="2026-01-21 11:28:44.424045258 +0000 UTC m=+1066.717074561" observedRunningTime="2026-01-21 11:28:44.625335391 +0000 UTC m=+1066.918364684" watchObservedRunningTime="2026-01-21 11:28:44.631148042 +0000 UTC m=+1066.924177333" Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.637226 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.637213428 podStartE2EDuration="2.637213428s" podCreationTimestamp="2026-01-21 11:28:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:44.636417998 +0000 UTC m=+1066.929447290" watchObservedRunningTime="2026-01-21 11:28:44.637213428 +0000 UTC m=+1066.930242720" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.199093 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.233844 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.245319 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.245490 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="dnsmasq-dns" containerID="cri-o://32fb68e8168ce229b5eee2857a8b6f19fa0565f1c232d572bde25ae83e70fb37" gracePeriod=10 Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.627075 4824 generic.go:334] "Generic (PLEG): container finished" podID="8247185f-987d-4f6e-b633-042907078e39" containerID="32fb68e8168ce229b5eee2857a8b6f19fa0565f1c232d572bde25ae83e70fb37" exitCode=0 Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.627290 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" event={"ID":"8247185f-987d-4f6e-b633-042907078e39","Type":"ContainerDied","Data":"32fb68e8168ce229b5eee2857a8b6f19fa0565f1c232d572bde25ae83e70fb37"} Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.627480 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" event={"ID":"8247185f-987d-4f6e-b633-042907078e39","Type":"ContainerDied","Data":"cee4d42d6e8f113c3a920c0a14aa2caa944d9803b04e59596b4d760c6b1d3e91"} Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.627566 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cee4d42d6e8f113c3a920c0a14aa2caa944d9803b04e59596b4d760c6b1d3e91" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.659485 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.745108 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.745919 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.746095 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kz5l\" (UniqueName: \"kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.746196 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.746283 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.746423 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.754136 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l" (OuterVolumeSpecName: "kube-api-access-9kz5l") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "kube-api-access-9kz5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.789709 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.791873 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.793064 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config" (OuterVolumeSpecName: "config") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.796752 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.798712 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848382 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848619 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848632 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848642 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kz5l\" (UniqueName: \"kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848651 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848659 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:47 crc kubenswrapper[4824]: I0121 11:28:47.633854 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:47 crc kubenswrapper[4824]: I0121 11:28:47.665486 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:47 crc kubenswrapper[4824]: I0121 11:28:47.672551 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:48 crc kubenswrapper[4824]: I0121 11:28:48.057627 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8247185f-987d-4f6e-b633-042907078e39" path="/var/lib/kubelet/pods/8247185f-987d-4f6e-b633-042907078e39/volumes" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.232002 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.246478 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.678245 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.786161 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-6c642"] Jan 21 11:28:51 crc kubenswrapper[4824]: E0121 11:28:51.786478 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="init" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.786494 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="init" Jan 21 11:28:51 crc kubenswrapper[4824]: E0121 11:28:51.786509 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="dnsmasq-dns" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.786516 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="dnsmasq-dns" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.786696 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="dnsmasq-dns" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.788713 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.795057 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6c642"] Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.795203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.795545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.930457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.930770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.930789 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.930853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcvgk\" (UniqueName: \"kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.032246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.032285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.032353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcvgk\" (UniqueName: \"kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.032461 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.037007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.037165 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.037882 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.044719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcvgk\" (UniqueName: \"kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.106321 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.470707 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6c642"] Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.673604 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6c642" event={"ID":"12c162ff-82c2-4166-8a44-92e0455cae39","Type":"ContainerStarted","Data":"7468f8bf7c17d51a36aabd9a1f3d9c19a0cb1f2b474c4bfcd676598418dd86f7"} Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.673652 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6c642" event={"ID":"12c162ff-82c2-4166-8a44-92e0455cae39","Type":"ContainerStarted","Data":"7a2f6e5181f30abf986b63626214d85313d1465ba029627b22f64bfd58a63370"} Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.690976 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-6c642" podStartSLOduration=1.690948538 podStartE2EDuration="1.690948538s" podCreationTimestamp="2026-01-21 11:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:52.685937899 +0000 UTC m=+1074.978967191" watchObservedRunningTime="2026-01-21 11:28:52.690948538 +0000 UTC m=+1074.983977829" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.950137 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.950174 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:28:53 crc kubenswrapper[4824]: I0121 11:28:53.965061 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:53 crc kubenswrapper[4824]: I0121 11:28:53.965094 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:56 crc kubenswrapper[4824]: I0121 11:28:56.702869 4824 generic.go:334] "Generic (PLEG): container finished" podID="12c162ff-82c2-4166-8a44-92e0455cae39" containerID="7468f8bf7c17d51a36aabd9a1f3d9c19a0cb1f2b474c4bfcd676598418dd86f7" exitCode=0 Jan 21 11:28:56 crc kubenswrapper[4824]: I0121 11:28:56.702975 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6c642" event={"ID":"12c162ff-82c2-4166-8a44-92e0455cae39","Type":"ContainerDied","Data":"7468f8bf7c17d51a36aabd9a1f3d9c19a0cb1f2b474c4bfcd676598418dd86f7"} Jan 21 11:28:57 crc kubenswrapper[4824]: I0121 11:28:57.930156 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.029602 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") pod \"12c162ff-82c2-4166-8a44-92e0455cae39\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.029665 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle\") pod \"12c162ff-82c2-4166-8a44-92e0455cae39\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.029693 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcvgk\" (UniqueName: \"kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk\") pod \"12c162ff-82c2-4166-8a44-92e0455cae39\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.029720 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts\") pod \"12c162ff-82c2-4166-8a44-92e0455cae39\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.033653 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts" (OuterVolumeSpecName: "scripts") pod "12c162ff-82c2-4166-8a44-92e0455cae39" (UID: "12c162ff-82c2-4166-8a44-92e0455cae39"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.033909 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk" (OuterVolumeSpecName: "kube-api-access-lcvgk") pod "12c162ff-82c2-4166-8a44-92e0455cae39" (UID: "12c162ff-82c2-4166-8a44-92e0455cae39"). InnerVolumeSpecName "kube-api-access-lcvgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:58 crc kubenswrapper[4824]: E0121 11:28:58.047430 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data podName:12c162ff-82c2-4166-8a44-92e0455cae39 nodeName:}" failed. No retries permitted until 2026-01-21 11:28:58.547413035 +0000 UTC m=+1080.840442327 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data") pod "12c162ff-82c2-4166-8a44-92e0455cae39" (UID: "12c162ff-82c2-4166-8a44-92e0455cae39") : error deleting /var/lib/kubelet/pods/12c162ff-82c2-4166-8a44-92e0455cae39/volume-subpaths: remove /var/lib/kubelet/pods/12c162ff-82c2-4166-8a44-92e0455cae39/volume-subpaths: no such file or directory Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.049260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12c162ff-82c2-4166-8a44-92e0455cae39" (UID: "12c162ff-82c2-4166-8a44-92e0455cae39"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.132138 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.132165 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcvgk\" (UniqueName: \"kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.132177 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.640848 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") pod \"12c162ff-82c2-4166-8a44-92e0455cae39\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.644742 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data" (OuterVolumeSpecName: "config-data") pod "12c162ff-82c2-4166-8a44-92e0455cae39" (UID: "12c162ff-82c2-4166-8a44-92e0455cae39"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.719855 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6c642" event={"ID":"12c162ff-82c2-4166-8a44-92e0455cae39","Type":"ContainerDied","Data":"7a2f6e5181f30abf986b63626214d85313d1465ba029627b22f64bfd58a63370"} Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.719888 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a2f6e5181f30abf986b63626214d85313d1465ba029627b22f64bfd58a63370" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.719916 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.743162 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.870597 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.870853 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-log" containerID="cri-o://4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218" gracePeriod=30 Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.870907 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-api" containerID="cri-o://ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331" gracePeriod=30 Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.879576 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.879743 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a8b9cd0d-c7af-455d-a170-1521870a07f6" containerName="nova-scheduler-scheduler" containerID="cri-o://be161c0756cab08e44d6825e362adf099303c412379798ba94c4f4bf705f4830" gracePeriod=30 Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.895901 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.896170 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" containerID="cri-o://55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9" gracePeriod=30 Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.896221 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" containerID="cri-o://12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c" gracePeriod=30 Jan 21 11:28:59 crc kubenswrapper[4824]: I0121 11:28:59.727633 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad597758-1f2c-4f84-8035-59fa20e68084" containerID="4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218" exitCode=143 Jan 21 11:28:59 crc kubenswrapper[4824]: I0121 11:28:59.727697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerDied","Data":"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218"} Jan 21 11:28:59 crc kubenswrapper[4824]: I0121 11:28:59.729731 4824 generic.go:334] "Generic (PLEG): container finished" podID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerID="55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9" exitCode=143 Jan 21 11:28:59 crc kubenswrapper[4824]: I0121 11:28:59.729757 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerDied","Data":"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9"} Jan 21 11:29:00 crc kubenswrapper[4824]: I0121 11:29:00.737367 4824 generic.go:334] "Generic (PLEG): container finished" podID="a8b9cd0d-c7af-455d-a170-1521870a07f6" containerID="be161c0756cab08e44d6825e362adf099303c412379798ba94c4f4bf705f4830" exitCode=0 Jan 21 11:29:00 crc kubenswrapper[4824]: I0121 11:29:00.737446 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8b9cd0d-c7af-455d-a170-1521870a07f6","Type":"ContainerDied","Data":"be161c0756cab08e44d6825e362adf099303c412379798ba94c4f4bf705f4830"} Jan 21 11:29:00 crc kubenswrapper[4824]: I0121 11:29:00.957234 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.078919 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data\") pod \"a8b9cd0d-c7af-455d-a170-1521870a07f6\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.079085 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsphm\" (UniqueName: \"kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm\") pod \"a8b9cd0d-c7af-455d-a170-1521870a07f6\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.079165 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle\") pod \"a8b9cd0d-c7af-455d-a170-1521870a07f6\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.083253 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm" (OuterVolumeSpecName: "kube-api-access-nsphm") pod "a8b9cd0d-c7af-455d-a170-1521870a07f6" (UID: "a8b9cd0d-c7af-455d-a170-1521870a07f6"). InnerVolumeSpecName "kube-api-access-nsphm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.099861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8b9cd0d-c7af-455d-a170-1521870a07f6" (UID: "a8b9cd0d-c7af-455d-a170-1521870a07f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.115359 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data" (OuterVolumeSpecName: "config-data") pod "a8b9cd0d-c7af-455d-a170-1521870a07f6" (UID: "a8b9cd0d-c7af-455d-a170-1521870a07f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.182794 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.183286 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsphm\" (UniqueName: \"kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.183318 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.744533 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8b9cd0d-c7af-455d-a170-1521870a07f6","Type":"ContainerDied","Data":"bcd965629c724abfbdcc4781034ea86d257b188520a919034221d6a8bf2c0f86"} Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.744583 4824 scope.go:117] "RemoveContainer" containerID="be161c0756cab08e44d6825e362adf099303c412379798ba94c4f4bf705f4830" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.744592 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.768257 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.776717 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.784451 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:29:01 crc kubenswrapper[4824]: E0121 11:29:01.784824 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c162ff-82c2-4166-8a44-92e0455cae39" containerName="nova-manage" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.784841 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c162ff-82c2-4166-8a44-92e0455cae39" containerName="nova-manage" Jan 21 11:29:01 crc kubenswrapper[4824]: E0121 11:29:01.784873 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8b9cd0d-c7af-455d-a170-1521870a07f6" containerName="nova-scheduler-scheduler" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.784881 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8b9cd0d-c7af-455d-a170-1521870a07f6" containerName="nova-scheduler-scheduler" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.785057 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c162ff-82c2-4166-8a44-92e0455cae39" containerName="nova-manage" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.785092 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8b9cd0d-c7af-455d-a170-1521870a07f6" containerName="nova-scheduler-scheduler" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.785637 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.787325 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.794382 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.894844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.895036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-config-data\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.895092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnbn7\" (UniqueName: \"kubernetes.io/projected/7860ffbb-b435-4736-b2d3-f8871e70dc7a-kube-api-access-lnbn7\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.996580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-config-data\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.997162 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnbn7\" (UniqueName: \"kubernetes.io/projected/7860ffbb-b435-4736-b2d3-f8871e70dc7a-kube-api-access-lnbn7\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.997241 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.001166 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-config-data\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.001450 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.010208 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnbn7\" (UniqueName: \"kubernetes.io/projected/7860ffbb-b435-4736-b2d3-f8871e70dc7a-kube-api-access-lnbn7\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.011226 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": dial tcp 10.217.0.204:8775: connect: connection refused" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.011229 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": dial tcp 10.217.0.204:8775: connect: connection refused" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.057664 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8b9cd0d-c7af-455d-a170-1521870a07f6" path="/var/lib/kubelet/pods/a8b9cd0d-c7af-455d-a170-1521870a07f6/volumes" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.101899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.398903 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.415153 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506032 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506114 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506267 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22h96\" (UniqueName: \"kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506293 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506922 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs" (OuterVolumeSpecName: "logs") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.509178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96" (OuterVolumeSpecName: "kube-api-access-22h96") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "kube-api-access-22h96". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.530539 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data" (OuterVolumeSpecName: "config-data") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.530612 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.536852 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: W0121 11:29:02.537448 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7860ffbb_b435_4736_b2d3_f8871e70dc7a.slice/crio-54b1920e6073c5b64a316ce4f6144c773b3d456175a5da934941d89af91d6ff5 WatchSource:0}: Error finding container 54b1920e6073c5b64a316ce4f6144c773b3d456175a5da934941d89af91d6ff5: Status 404 returned error can't find the container with id 54b1920e6073c5b64a316ce4f6144c773b3d456175a5da934941d89af91d6ff5 Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.546631 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.550309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.608065 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpvwd\" (UniqueName: \"kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd\") pod \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.608098 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data\") pod \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.608151 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs\") pod \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.608244 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs\") pod \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.608982 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs" (OuterVolumeSpecName: "logs") pod "3b9162ed-f3d2-4877-b8a3-6342a25200b7" (UID: "3b9162ed-f3d2-4877-b8a3-6342a25200b7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.609115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle\") pod \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.609947 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22h96\" (UniqueName: \"kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.609992 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610009 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610018 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610025 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610034 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610103 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd" (OuterVolumeSpecName: "kube-api-access-lpvwd") pod "3b9162ed-f3d2-4877-b8a3-6342a25200b7" (UID: "3b9162ed-f3d2-4877-b8a3-6342a25200b7"). InnerVolumeSpecName "kube-api-access-lpvwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.626675 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data" (OuterVolumeSpecName: "config-data") pod "3b9162ed-f3d2-4877-b8a3-6342a25200b7" (UID: "3b9162ed-f3d2-4877-b8a3-6342a25200b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.628363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b9162ed-f3d2-4877-b8a3-6342a25200b7" (UID: "3b9162ed-f3d2-4877-b8a3-6342a25200b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.650822 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3b9162ed-f3d2-4877-b8a3-6342a25200b7" (UID: "3b9162ed-f3d2-4877-b8a3-6342a25200b7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.712111 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.712140 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpvwd\" (UniqueName: \"kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.712151 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.712159 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.752864 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad597758-1f2c-4f84-8035-59fa20e68084" containerID="ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331" exitCode=0 Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.752922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerDied","Data":"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.752975 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerDied","Data":"04682cea27407b0d473ba48a7b6b0c26d36ec7ff8680fb81116bbb44f989542f"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.752993 4824 scope.go:117] "RemoveContainer" containerID="ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.753702 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.756571 4824 generic.go:334] "Generic (PLEG): container finished" podID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerID="12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c" exitCode=0 Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.756617 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.756631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerDied","Data":"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.756655 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerDied","Data":"cf896b02cb7f72cd22b94afd5cb645eb3dc94f2424b20aa748a3b1e35979c8a8"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.758646 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7860ffbb-b435-4736-b2d3-f8871e70dc7a","Type":"ContainerStarted","Data":"92316441f6f75de082981736928c3af2413970490f85031826e288c70a591cca"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.758671 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7860ffbb-b435-4736-b2d3-f8871e70dc7a","Type":"ContainerStarted","Data":"54b1920e6073c5b64a316ce4f6144c773b3d456175a5da934941d89af91d6ff5"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.769447 4824 scope.go:117] "RemoveContainer" containerID="4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.770915 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.7709049430000001 podStartE2EDuration="1.770904943s" podCreationTimestamp="2026-01-21 11:29:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:29:02.768465102 +0000 UTC m=+1085.061494394" watchObservedRunningTime="2026-01-21 11:29:02.770904943 +0000 UTC m=+1085.063934235" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.785667 4824 scope.go:117] "RemoveContainer" containerID="ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.786389 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331\": container with ID starting with ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331 not found: ID does not exist" containerID="ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.786422 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331"} err="failed to get container status \"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331\": rpc error: code = NotFound desc = could not find container \"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331\": container with ID starting with ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331 not found: ID does not exist" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.786442 4824 scope.go:117] "RemoveContainer" containerID="4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.786713 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218\": container with ID starting with 4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218 not found: ID does not exist" containerID="4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.786788 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218"} err="failed to get container status \"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218\": rpc error: code = NotFound desc = could not find container \"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218\": container with ID starting with 4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218 not found: ID does not exist" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.786877 4824 scope.go:117] "RemoveContainer" containerID="12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.787302 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.797520 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.810668 4824 scope.go:117] "RemoveContainer" containerID="55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.810830 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.826146 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.839565 4824 scope.go:117] "RemoveContainer" containerID="12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.839663 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840045 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-api" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840062 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-api" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840084 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840090 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840096 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-log" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840101 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-log" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840115 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840121 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840119 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c\": container with ID starting with 12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c not found: ID does not exist" containerID="12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840154 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c"} err="failed to get container status \"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c\": rpc error: code = NotFound desc = could not find container \"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c\": container with ID starting with 12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c not found: ID does not exist" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840176 4824 scope.go:117] "RemoveContainer" containerID="55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840282 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-api" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840303 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840315 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840321 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-log" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840652 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9\": container with ID starting with 55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9 not found: ID does not exist" containerID="55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840672 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9"} err="failed to get container status \"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9\": rpc error: code = NotFound desc = could not find container \"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9\": container with ID starting with 55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9 not found: ID does not exist" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.841192 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.842891 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.843085 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.844082 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.845995 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.852534 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.853867 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.857250 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.857399 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.858472 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018152 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018243 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/654a41d0-b44a-4194-a514-a3c60126d0d9-logs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018754 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-config-data\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018994 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.019022 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.019043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6vg4\" (UniqueName: \"kubernetes.io/projected/654a41d0-b44a-4194-a514-a3c60126d0d9-kube-api-access-f6vg4\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.019057 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-config-data\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.019077 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55525bd3-ae2b-494d-9a33-dd7d00c576b2-logs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.019205 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96dc4\" (UniqueName: \"kubernetes.io/projected/55525bd3-ae2b-494d-9a33-dd7d00c576b2-kube-api-access-96dc4\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-config-data\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121521 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6vg4\" (UniqueName: \"kubernetes.io/projected/654a41d0-b44a-4194-a514-a3c60126d0d9-kube-api-access-f6vg4\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55525bd3-ae2b-494d-9a33-dd7d00c576b2-logs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121569 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96dc4\" (UniqueName: \"kubernetes.io/projected/55525bd3-ae2b-494d-9a33-dd7d00c576b2-kube-api-access-96dc4\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121779 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/654a41d0-b44a-4194-a514-a3c60126d0d9-logs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.122265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55525bd3-ae2b-494d-9a33-dd7d00c576b2-logs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121840 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.122395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.122455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-config-data\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.122553 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/654a41d0-b44a-4194-a514-a3c60126d0d9-logs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.132441 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-config-data\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.139129 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.140898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.141057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.141163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-config-data\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.141281 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.142695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96dc4\" (UniqueName: \"kubernetes.io/projected/55525bd3-ae2b-494d-9a33-dd7d00c576b2-kube-api-access-96dc4\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.143817 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6vg4\" (UniqueName: \"kubernetes.io/projected/654a41d0-b44a-4194-a514-a3c60126d0d9-kube-api-access-f6vg4\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.151753 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.157396 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.165157 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.531105 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.600812 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:29:03 crc kubenswrapper[4824]: W0121 11:29:03.601781 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55525bd3_ae2b_494d_9a33_dd7d00c576b2.slice/crio-450475ff361db096670a972124133a11e8549d19be646c798d0b7d4f12c00208 WatchSource:0}: Error finding container 450475ff361db096670a972124133a11e8549d19be646c798d0b7d4f12c00208: Status 404 returned error can't find the container with id 450475ff361db096670a972124133a11e8549d19be646c798d0b7d4f12c00208 Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.767538 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"654a41d0-b44a-4194-a514-a3c60126d0d9","Type":"ContainerStarted","Data":"39d8671f823e852b86d992e4469bf9ba938c5889950d2d64d5b2f68626370183"} Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.767752 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"654a41d0-b44a-4194-a514-a3c60126d0d9","Type":"ContainerStarted","Data":"c9941457bd25dfdb724c33502c3621bf31db3922f63f60e11fce2ece4cef0c7f"} Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.767766 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"654a41d0-b44a-4194-a514-a3c60126d0d9","Type":"ContainerStarted","Data":"750307738ca25e9c7bf090aaadeeb9fbd671e51a467f1d767a28cf93a8e85a15"} Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.777325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55525bd3-ae2b-494d-9a33-dd7d00c576b2","Type":"ContainerStarted","Data":"7ccf6dca71329f80f6672211c8a36a53367d5588455ddf171422d231a8c4f4b8"} Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.777355 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55525bd3-ae2b-494d-9a33-dd7d00c576b2","Type":"ContainerStarted","Data":"450475ff361db096670a972124133a11e8549d19be646c798d0b7d4f12c00208"} Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.788456 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.788441911 podStartE2EDuration="1.788441911s" podCreationTimestamp="2026-01-21 11:29:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:29:03.782496571 +0000 UTC m=+1086.075525864" watchObservedRunningTime="2026-01-21 11:29:03.788441911 +0000 UTC m=+1086.081471204" Jan 21 11:29:04 crc kubenswrapper[4824]: I0121 11:29:04.061083 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" path="/var/lib/kubelet/pods/3b9162ed-f3d2-4877-b8a3-6342a25200b7/volumes" Jan 21 11:29:04 crc kubenswrapper[4824]: I0121 11:29:04.061667 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" path="/var/lib/kubelet/pods/ad597758-1f2c-4f84-8035-59fa20e68084/volumes" Jan 21 11:29:04 crc kubenswrapper[4824]: I0121 11:29:04.794376 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55525bd3-ae2b-494d-9a33-dd7d00c576b2","Type":"ContainerStarted","Data":"58fb9a1b8b1e252d070cfa3d6bb163d38f884129a6fe386c671924e7aca82b98"} Jan 21 11:29:04 crc kubenswrapper[4824]: I0121 11:29:04.807889 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8078750550000002 podStartE2EDuration="2.807875055s" podCreationTimestamp="2026-01-21 11:29:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:29:04.80708257 +0000 UTC m=+1087.100111862" watchObservedRunningTime="2026-01-21 11:29:04.807875055 +0000 UTC m=+1087.100904347" Jan 21 11:29:07 crc kubenswrapper[4824]: I0121 11:29:07.102400 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 21 11:29:08 crc kubenswrapper[4824]: I0121 11:29:08.165207 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 21 11:29:08 crc kubenswrapper[4824]: I0121 11:29:08.165433 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 21 11:29:10 crc kubenswrapper[4824]: I0121 11:29:10.928467 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 21 11:29:12 crc kubenswrapper[4824]: I0121 11:29:12.102656 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 21 11:29:12 crc kubenswrapper[4824]: I0121 11:29:12.122124 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 21 11:29:12 crc kubenswrapper[4824]: I0121 11:29:12.871374 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.158249 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.158466 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.165912 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.165943 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.720688 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.720872 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="2e3a943c-ee06-4e54-a984-34ecb1320f89" containerName="kube-state-metrics" containerID="cri-o://6e0a2a51cb10bfe8bc65ec0123c4fa76489cf9873a673344ebc1fbf8065155d5" gracePeriod=30 Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.864050 4824 generic.go:334] "Generic (PLEG): container finished" podID="2e3a943c-ee06-4e54-a984-34ecb1320f89" containerID="6e0a2a51cb10bfe8bc65ec0123c4fa76489cf9873a673344ebc1fbf8065155d5" exitCode=2 Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.864754 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2e3a943c-ee06-4e54-a984-34ecb1320f89","Type":"ContainerDied","Data":"6e0a2a51cb10bfe8bc65ec0123c4fa76489cf9873a673344ebc1fbf8065155d5"} Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.122973 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.168105 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="55525bd3-ae2b-494d-9a33-dd7d00c576b2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.168129 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="55525bd3-ae2b-494d-9a33-dd7d00c576b2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.182347 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="654a41d0-b44a-4194-a514-a3c60126d0d9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.182505 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="654a41d0-b44a-4194-a514-a3c60126d0d9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.199651 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6ndv\" (UniqueName: \"kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv\") pod \"2e3a943c-ee06-4e54-a984-34ecb1320f89\" (UID: \"2e3a943c-ee06-4e54-a984-34ecb1320f89\") " Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.206180 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv" (OuterVolumeSpecName: "kube-api-access-n6ndv") pod "2e3a943c-ee06-4e54-a984-34ecb1320f89" (UID: "2e3a943c-ee06-4e54-a984-34ecb1320f89"). InnerVolumeSpecName "kube-api-access-n6ndv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.302069 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6ndv\" (UniqueName: \"kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.871876 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2e3a943c-ee06-4e54-a984-34ecb1320f89","Type":"ContainerDied","Data":"6cb5ad5332507e249a47db656294541a580560b8688929dae90a9899b846d8f8"} Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.871916 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.871924 4824 scope.go:117] "RemoveContainer" containerID="6e0a2a51cb10bfe8bc65ec0123c4fa76489cf9873a673344ebc1fbf8065155d5" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.896762 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.940525 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.950489 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:14 crc kubenswrapper[4824]: E0121 11:29:14.950881 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e3a943c-ee06-4e54-a984-34ecb1320f89" containerName="kube-state-metrics" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.950898 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e3a943c-ee06-4e54-a984-34ecb1320f89" containerName="kube-state-metrics" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.951114 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e3a943c-ee06-4e54-a984-34ecb1320f89" containerName="kube-state-metrics" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.951705 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.953385 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.953592 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.972010 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.020450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.020528 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r25fc\" (UniqueName: \"kubernetes.io/projected/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-api-access-r25fc\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.020558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.020779 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.122483 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.122723 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.122765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r25fc\" (UniqueName: \"kubernetes.io/projected/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-api-access-r25fc\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.122789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.130295 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.131313 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.141497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r25fc\" (UniqueName: \"kubernetes.io/projected/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-api-access-r25fc\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.141517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.267994 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.291732 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.291971 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-central-agent" containerID="cri-o://f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f" gracePeriod=30 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.292030 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="sg-core" containerID="cri-o://ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a" gracePeriod=30 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.292057 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="proxy-httpd" containerID="cri-o://0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74" gracePeriod=30 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.292037 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-notification-agent" containerID="cri-o://ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253" gracePeriod=30 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.664557 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:15 crc kubenswrapper[4824]: W0121 11:29:15.672113 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd42233a9_cfd9_486d_b314_f17d6a9d3b03.slice/crio-bbba9f8a9650e396c656bdb12251ced4c13ba8643f286ba15c3458435cec5ce0 WatchSource:0}: Error finding container bbba9f8a9650e396c656bdb12251ced4c13ba8643f286ba15c3458435cec5ce0: Status 404 returned error can't find the container with id bbba9f8a9650e396c656bdb12251ced4c13ba8643f286ba15c3458435cec5ce0 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.879089 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d42233a9-cfd9-486d-b314-f17d6a9d3b03","Type":"ContainerStarted","Data":"bbba9f8a9650e396c656bdb12251ced4c13ba8643f286ba15c3458435cec5ce0"} Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882123 4824 generic.go:334] "Generic (PLEG): container finished" podID="55916065-772b-45ee-8871-37c229777deb" containerID="0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74" exitCode=0 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882148 4824 generic.go:334] "Generic (PLEG): container finished" podID="55916065-772b-45ee-8871-37c229777deb" containerID="ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a" exitCode=2 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882157 4824 generic.go:334] "Generic (PLEG): container finished" podID="55916065-772b-45ee-8871-37c229777deb" containerID="f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f" exitCode=0 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882170 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerDied","Data":"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74"} Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerDied","Data":"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a"} Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882194 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerDied","Data":"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f"} Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.057636 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e3a943c-ee06-4e54-a984-34ecb1320f89" path="/var/lib/kubelet/pods/2e3a943c-ee06-4e54-a984-34ecb1320f89/volumes" Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.064835 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.064879 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.894371 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d42233a9-cfd9-486d-b314-f17d6a9d3b03","Type":"ContainerStarted","Data":"29ffe115de3536006b5bf65e63db5d06e9c710a58c3108031b8df7712c9cdfe7"} Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.894700 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.913200 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.646854743 podStartE2EDuration="2.913186132s" podCreationTimestamp="2026-01-21 11:29:14 +0000 UTC" firstStartedPulling="2026-01-21 11:29:15.674439131 +0000 UTC m=+1097.967468423" lastFinishedPulling="2026-01-21 11:29:15.94077052 +0000 UTC m=+1098.233799812" observedRunningTime="2026-01-21 11:29:16.906544779 +0000 UTC m=+1099.199574071" watchObservedRunningTime="2026-01-21 11:29:16.913186132 +0000 UTC m=+1099.206215424" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.376196 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.464786 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.464851 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.464874 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.464895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.464939 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.465036 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvljc\" (UniqueName: \"kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.465072 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.466066 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.466177 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.469700 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc" (OuterVolumeSpecName: "kube-api-access-gvljc") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "kube-api-access-gvljc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.473534 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts" (OuterVolumeSpecName: "scripts") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.486299 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.518433 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.536125 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data" (OuterVolumeSpecName: "config-data") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567461 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvljc\" (UniqueName: \"kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567487 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567496 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567503 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567510 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567517 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567526 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.903116 4824 generic.go:334] "Generic (PLEG): container finished" podID="55916065-772b-45ee-8871-37c229777deb" containerID="ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253" exitCode=0 Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.903172 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.903198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerDied","Data":"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253"} Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.903444 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerDied","Data":"65e4d9ce199d66a5d4b888b55edee4667e5ba88065fc68ccd26d1ad79dbffbb0"} Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.903467 4824 scope.go:117] "RemoveContainer" containerID="0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.918097 4824 scope.go:117] "RemoveContainer" containerID="ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.929905 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.934726 4824 scope.go:117] "RemoveContainer" containerID="ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.939035 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951045 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.951462 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-central-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951496 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-central-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.951514 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="sg-core" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951521 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="sg-core" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.951535 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="proxy-httpd" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951540 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="proxy-httpd" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.951581 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-notification-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951588 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-notification-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951778 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-central-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951825 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="proxy-httpd" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951834 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-notification-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951852 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="sg-core" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.954550 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.956498 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.956852 4824 scope.go:117] "RemoveContainer" containerID="f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.956866 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.956921 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.956997 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.972910 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-scripts\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-config-data\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-run-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973293 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-log-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjpbm\" (UniqueName: \"kubernetes.io/projected/adc98d07-2b32-4f30-bdb7-40923468389e-kube-api-access-qjpbm\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973864 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.980251 4824 scope.go:117] "RemoveContainer" containerID="0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.980528 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74\": container with ID starting with 0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74 not found: ID does not exist" containerID="0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.980558 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74"} err="failed to get container status \"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74\": rpc error: code = NotFound desc = could not find container \"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74\": container with ID starting with 0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74 not found: ID does not exist" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.980575 4824 scope.go:117] "RemoveContainer" containerID="ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.980918 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a\": container with ID starting with ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a not found: ID does not exist" containerID="ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.980940 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a"} err="failed to get container status \"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a\": rpc error: code = NotFound desc = could not find container \"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a\": container with ID starting with ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a not found: ID does not exist" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.980967 4824 scope.go:117] "RemoveContainer" containerID="ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.981265 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253\": container with ID starting with ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253 not found: ID does not exist" containerID="ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.981285 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253"} err="failed to get container status \"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253\": rpc error: code = NotFound desc = could not find container \"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253\": container with ID starting with ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253 not found: ID does not exist" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.981297 4824 scope.go:117] "RemoveContainer" containerID="f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.981463 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f\": container with ID starting with f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f not found: ID does not exist" containerID="f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.981481 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f"} err="failed to get container status \"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f\": rpc error: code = NotFound desc = could not find container \"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f\": container with ID starting with f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f not found: ID does not exist" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.063261 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55916065-772b-45ee-8871-37c229777deb" path="/var/lib/kubelet/pods/55916065-772b-45ee-8871-37c229777deb/volumes" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.074824 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-run-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.074884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.074932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-log-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.074980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.075044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjpbm\" (UniqueName: \"kubernetes.io/projected/adc98d07-2b32-4f30-bdb7-40923468389e-kube-api-access-qjpbm\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.075104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.075124 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-scripts\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.075138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-config-data\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.076209 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-log-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.076641 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-run-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.078484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.078633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-config-data\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.078855 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.079401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-scripts\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.080033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.087645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjpbm\" (UniqueName: \"kubernetes.io/projected/adc98d07-2b32-4f30-bdb7-40923468389e-kube-api-access-qjpbm\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.277654 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.646710 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:18 crc kubenswrapper[4824]: W0121 11:29:18.654249 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadc98d07_2b32_4f30_bdb7_40923468389e.slice/crio-2916545b5864c587e6e609faa9526231fc3891c256061f4fb6e3de6ffb82750f WatchSource:0}: Error finding container 2916545b5864c587e6e609faa9526231fc3891c256061f4fb6e3de6ffb82750f: Status 404 returned error can't find the container with id 2916545b5864c587e6e609faa9526231fc3891c256061f4fb6e3de6ffb82750f Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.911800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"adc98d07-2b32-4f30-bdb7-40923468389e","Type":"ContainerStarted","Data":"2916545b5864c587e6e609faa9526231fc3891c256061f4fb6e3de6ffb82750f"} Jan 21 11:29:19 crc kubenswrapper[4824]: I0121 11:29:19.918978 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"adc98d07-2b32-4f30-bdb7-40923468389e","Type":"ContainerStarted","Data":"a90e401f791539feac129f45795db65909d69b5d726e41e25c99fb59745fa42e"} Jan 21 11:29:20 crc kubenswrapper[4824]: I0121 11:29:20.929678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"adc98d07-2b32-4f30-bdb7-40923468389e","Type":"ContainerStarted","Data":"90a1264a92ff9f804e52949d65ec8a022c71f5ff93ed7e2a3db6674d2d1d512c"} Jan 21 11:29:20 crc kubenswrapper[4824]: I0121 11:29:20.930152 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"adc98d07-2b32-4f30-bdb7-40923468389e","Type":"ContainerStarted","Data":"b03e582376e52022beb2bcf9dc6dfe2b49f9da87daff5d1c7851a1edc48f84cb"} Jan 21 11:29:22 crc kubenswrapper[4824]: I0121 11:29:22.945177 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"adc98d07-2b32-4f30-bdb7-40923468389e","Type":"ContainerStarted","Data":"3d1045aabdaf7cb011d4cf495207121285049e47b5e1472749fd1020f5ee11e9"} Jan 21 11:29:22 crc kubenswrapper[4824]: I0121 11:29:22.945575 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:29:22 crc kubenswrapper[4824]: I0121 11:29:22.966065 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.75376542 podStartE2EDuration="5.966050794s" podCreationTimestamp="2026-01-21 11:29:17 +0000 UTC" firstStartedPulling="2026-01-21 11:29:18.656472743 +0000 UTC m=+1100.949502035" lastFinishedPulling="2026-01-21 11:29:21.868758117 +0000 UTC m=+1104.161787409" observedRunningTime="2026-01-21 11:29:22.95960423 +0000 UTC m=+1105.252633532" watchObservedRunningTime="2026-01-21 11:29:22.966050794 +0000 UTC m=+1105.259080087" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.163089 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.163404 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.163555 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.166889 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.173208 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.178794 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.180554 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.953816 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.956628 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.958397 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 21 11:29:25 crc kubenswrapper[4824]: I0121 11:29:25.277569 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 21 11:29:46 crc kubenswrapper[4824]: I0121 11:29:46.065175 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:29:46 crc kubenswrapper[4824]: I0121 11:29:46.065487 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:29:48 crc kubenswrapper[4824]: I0121 11:29:48.285195 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 21 11:29:55 crc kubenswrapper[4824]: I0121 11:29:55.955165 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:29:56 crc kubenswrapper[4824]: I0121 11:29:56.679703 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:29:59 crc kubenswrapper[4824]: I0121 11:29:59.347089 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="rabbitmq" containerID="cri-o://407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7" gracePeriod=604797 Jan 21 11:29:59 crc kubenswrapper[4824]: I0121 11:29:59.851710 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="rabbitmq" containerID="cri-o://902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32" gracePeriod=604797 Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.146395 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz"] Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.147698 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.149672 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.150616 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.153098 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz"] Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.282466 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.282942 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6hhk\" (UniqueName: \"kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.283201 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.384677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6hhk\" (UniqueName: \"kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.385382 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.385528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.386290 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.390373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.398786 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6hhk\" (UniqueName: \"kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.473908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.854975 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz"] Jan 21 11:30:01 crc kubenswrapper[4824]: I0121 11:30:01.196294 4824 generic.go:334] "Generic (PLEG): container finished" podID="4c02bf47-f856-44f3-af2c-cc96c45471c1" containerID="dd264d8d4b02d26df3192da84e03b87f54958338166fb00a9d9562f0fbc98220" exitCode=0 Jan 21 11:30:01 crc kubenswrapper[4824]: I0121 11:30:01.196511 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" event={"ID":"4c02bf47-f856-44f3-af2c-cc96c45471c1","Type":"ContainerDied","Data":"dd264d8d4b02d26df3192da84e03b87f54958338166fb00a9d9562f0fbc98220"} Jan 21 11:30:01 crc kubenswrapper[4824]: I0121 11:30:01.196535 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" event={"ID":"4c02bf47-f856-44f3-af2c-cc96c45471c1","Type":"ContainerStarted","Data":"5404443892f777ddba4e33e0fe95bb8c01a0c937d229122cc30c76b5b89517d6"} Jan 21 11:30:01 crc kubenswrapper[4824]: I0121 11:30:01.373268 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Jan 21 11:30:01 crc kubenswrapper[4824]: I0121 11:30:01.631637 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.454599 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.520845 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume\") pod \"4c02bf47-f856-44f3-af2c-cc96c45471c1\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.520969 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6hhk\" (UniqueName: \"kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk\") pod \"4c02bf47-f856-44f3-af2c-cc96c45471c1\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.521015 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume\") pod \"4c02bf47-f856-44f3-af2c-cc96c45471c1\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.521507 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume" (OuterVolumeSpecName: "config-volume") pod "4c02bf47-f856-44f3-af2c-cc96c45471c1" (UID: "4c02bf47-f856-44f3-af2c-cc96c45471c1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.525468 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4c02bf47-f856-44f3-af2c-cc96c45471c1" (UID: "4c02bf47-f856-44f3-af2c-cc96c45471c1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.525517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk" (OuterVolumeSpecName: "kube-api-access-h6hhk") pod "4c02bf47-f856-44f3-af2c-cc96c45471c1" (UID: "4c02bf47-f856-44f3-af2c-cc96c45471c1"). InnerVolumeSpecName "kube-api-access-h6hhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.623177 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.623207 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6hhk\" (UniqueName: \"kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.623220 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:03 crc kubenswrapper[4824]: I0121 11:30:03.209866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" event={"ID":"4c02bf47-f856-44f3-af2c-cc96c45471c1","Type":"ContainerDied","Data":"5404443892f777ddba4e33e0fe95bb8c01a0c937d229122cc30c76b5b89517d6"} Jan 21 11:30:03 crc kubenswrapper[4824]: I0121 11:30:03.209900 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:03 crc kubenswrapper[4824]: I0121 11:30:03.209907 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5404443892f777ddba4e33e0fe95bb8c01a0c937d229122cc30c76b5b89517d6" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.713433 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876569 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876715 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876744 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876806 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876841 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876862 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876909 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876949 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkwkl\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.877371 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.877606 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.877738 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.878139 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.878163 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.878174 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.881065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.881563 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info" (OuterVolumeSpecName: "pod-info") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.881572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl" (OuterVolumeSpecName: "kube-api-access-zkwkl") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "kube-api-access-zkwkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.881634 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.884074 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.914340 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data" (OuterVolumeSpecName: "config-data") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.928004 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf" (OuterVolumeSpecName: "server-conf") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.972927 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979802 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979844 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979855 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkwkl\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979864 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979872 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979878 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979885 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979893 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:05.999774 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.082520 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.178906 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.234201 4824 generic.go:334] "Generic (PLEG): container finished" podID="156fec50-b486-4e84-a7bf-b40491a863c7" containerID="902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32" exitCode=0 Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.234246 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.234272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerDied","Data":"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32"} Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.234298 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerDied","Data":"37d36a29724a133bc6f138979451aa8f9c0eeec9cad658937121a853dcbd295e"} Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.234315 4824 scope.go:117] "RemoveContainer" containerID="902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.236427 4824 generic.go:334] "Generic (PLEG): container finished" podID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerID="407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7" exitCode=0 Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.236458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerDied","Data":"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7"} Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.236477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerDied","Data":"8ee895c662c7814313d66afa49854eb939a2dc25462867e35d0362794391ee25"} Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.236524 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.251928 4824 scope.go:117] "RemoveContainer" containerID="57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.265868 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.277023 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.282107 4824 scope.go:117] "RemoveContainer" containerID="902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.284928 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32\": container with ID starting with 902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32 not found: ID does not exist" containerID="902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.284973 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9nc5\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.284984 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32"} err="failed to get container status \"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32\": rpc error: code = NotFound desc = could not find container \"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32\": container with ID starting with 902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32 not found: ID does not exist" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285010 4824 scope.go:117] "RemoveContainer" containerID="57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285021 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285082 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285167 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285210 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285267 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285314 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285354 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.285360 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487\": container with ID starting with 57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487 not found: ID does not exist" containerID="57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285393 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487"} err="failed to get container status \"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487\": rpc error: code = NotFound desc = could not find container \"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487\": container with ID starting with 57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487 not found: ID does not exist" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285416 4824 scope.go:117] "RemoveContainer" containerID="407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285606 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285891 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.286150 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.286975 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.286993 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.287002 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.293875 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.294396 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5" (OuterVolumeSpecName: "kube-api-access-j9nc5") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "kube-api-access-j9nc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.294440 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.294645 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.294760 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.295107 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="setup-container" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295124 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="setup-container" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.295140 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295147 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.295159 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c02bf47-f856-44f3-af2c-cc96c45471c1" containerName="collect-profiles" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295165 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c02bf47-f856-44f3-af2c-cc96c45471c1" containerName="collect-profiles" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.295176 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.295188 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="setup-container" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295193 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="setup-container" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295369 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c02bf47-f856-44f3-af2c-cc96c45471c1" containerName="collect-profiles" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295384 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295398 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.296750 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301405 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301447 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301597 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301612 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301692 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301758 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-22jwj" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301806 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.303174 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info" (OuterVolumeSpecName: "pod-info") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.307657 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.308591 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data" (OuterVolumeSpecName: "config-data") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.323787 4824 scope.go:117] "RemoveContainer" containerID="12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.343511 4824 scope.go:117] "RemoveContainer" containerID="407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.343968 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7\": container with ID starting with 407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7 not found: ID does not exist" containerID="407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.343996 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7"} err="failed to get container status \"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7\": rpc error: code = NotFound desc = could not find container \"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7\": container with ID starting with 407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7 not found: ID does not exist" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.344039 4824 scope.go:117] "RemoveContainer" containerID="12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.344501 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17\": container with ID starting with 12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17 not found: ID does not exist" containerID="12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.344542 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17"} err="failed to get container status \"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17\": rpc error: code = NotFound desc = could not find container \"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17\": container with ID starting with 12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17 not found: ID does not exist" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.357111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf" (OuterVolumeSpecName: "server-conf") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.387918 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.387981 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388010 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4a4d725-1a49-4292-af74-6ab09ab15cdb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388079 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388120 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4a4d725-1a49-4292-af74-6ab09ab15cdb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388142 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388203 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388220 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388300 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8m7b\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-kube-api-access-f8m7b\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388599 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388619 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388629 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388639 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388647 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388674 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388683 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9nc5\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.392606 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.404820 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490360 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490451 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4a4d725-1a49-4292-af74-6ab09ab15cdb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4a4d725-1a49-4292-af74-6ab09ab15cdb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.491241 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.491871 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.491067 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.491880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492224 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492360 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492440 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492563 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492574 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492849 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8m7b\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-kube-api-access-f8m7b\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.493303 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.493355 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.493362 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.493510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4a4d725-1a49-4292-af74-6ab09ab15cdb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.493904 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.494040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4a4d725-1a49-4292-af74-6ab09ab15cdb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.494585 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.507925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8m7b\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-kube-api-access-f8m7b\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.519175 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.615442 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.619632 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.626142 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.639520 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.641482 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.644479 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.644743 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.644970 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.645201 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zjdwv" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.645429 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.645597 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.645842 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.680811 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799098 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799293 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vbvn\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-kube-api-access-7vbvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799539 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799684 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799916 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.800032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.800058 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.901788 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.901878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.901984 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902007 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902052 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902137 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vbvn\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-kube-api-access-7vbvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902250 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902543 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.903327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.903728 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.903807 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.903951 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.904228 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.908986 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.909171 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.909289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.911534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.918126 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vbvn\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-kube-api-access-7vbvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.926582 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.011187 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.032331 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.179265 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.185046 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.188742 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.189157 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.245461 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4a4d725-1a49-4292-af74-6ab09ab15cdb","Type":"ContainerStarted","Data":"8b84f8fc3aaf1e628cbd83d5b25c754b6f48c97cb5b78b8c4f181d0a7a03affa"} Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309513 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309605 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309629 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309774 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txbf7\" (UniqueName: \"kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.310004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411692 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411787 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412035 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txbf7\" (UniqueName: \"kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412659 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412735 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412838 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.413222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.426112 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txbf7\" (UniqueName: \"kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.499556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.902511 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.910525 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.057225 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" path="/var/lib/kubelet/pods/156fec50-b486-4e84-a7bf-b40491a863c7/volumes" Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.059178 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" path="/var/lib/kubelet/pods/84e74d7b-18a0-4a3f-8680-6246ac538a6e/volumes" Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.281107 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4a4d725-1a49-4292-af74-6ab09ab15cdb","Type":"ContainerStarted","Data":"8915f56cc0770d64ee6a0d4ce2d2834b7ba05053d7b8c9baa030f6a50a88dee1"} Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.292103 4824 generic.go:334] "Generic (PLEG): container finished" podID="aed87a15-74e4-47e3-8f90-f98672d62792" containerID="2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de" exitCode=0 Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.292153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" event={"ID":"aed87a15-74e4-47e3-8f90-f98672d62792","Type":"ContainerDied","Data":"2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de"} Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.292172 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" event={"ID":"aed87a15-74e4-47e3-8f90-f98672d62792","Type":"ContainerStarted","Data":"d3d7d6a8cb817cd1fffab43e3dc0bcf2917c60a0925857f2eded916d57a51538"} Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.301111 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762","Type":"ContainerStarted","Data":"cb0628a064449b01379611bee44c81428d76b736554608f5c0a5f9e831d04f4b"} Jan 21 11:30:09 crc kubenswrapper[4824]: I0121 11:30:09.309076 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" event={"ID":"aed87a15-74e4-47e3-8f90-f98672d62792","Type":"ContainerStarted","Data":"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f"} Jan 21 11:30:09 crc kubenswrapper[4824]: I0121 11:30:09.309429 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:09 crc kubenswrapper[4824]: I0121 11:30:09.310095 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762","Type":"ContainerStarted","Data":"1ffd78c66a880f9e4de61c175a16fa8e25a34ed2dd520861ea49b68a2f4172fe"} Jan 21 11:30:09 crc kubenswrapper[4824]: I0121 11:30:09.324673 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" podStartSLOduration=2.324658385 podStartE2EDuration="2.324658385s" podCreationTimestamp="2026-01-21 11:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:30:09.321753348 +0000 UTC m=+1151.614782690" watchObservedRunningTime="2026-01-21 11:30:09.324658385 +0000 UTC m=+1151.617687677" Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.065321 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.065722 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.065762 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.066405 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.066456 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690" gracePeriod=600 Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.358715 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690" exitCode=0 Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.358867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690"} Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.358993 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4"} Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.359016 4824 scope.go:117] "RemoveContainer" containerID="365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.501094 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.537450 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.537669 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="dnsmasq-dns" containerID="cri-o://8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575" gracePeriod=10 Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.650026 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-26d8b"] Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.651595 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.677591 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-26d8b"] Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.786862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787458 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvbf7\" (UniqueName: \"kubernetes.io/projected/976a6446-e79c-42db-be45-43147767f1da-kube-api-access-pvbf7\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787503 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-config\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787557 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787654 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787779 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889777 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvbf7\" (UniqueName: \"kubernetes.io/projected/976a6446-e79c-42db-be45-43147767f1da-kube-api-access-pvbf7\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-config\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889892 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.890057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.891186 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.891189 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.891192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-config\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.891283 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.891734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.892138 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.907445 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvbf7\" (UniqueName: \"kubernetes.io/projected/976a6446-e79c-42db-be45-43147767f1da-kube-api-access-pvbf7\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.996738 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.002434 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.092504 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.093059 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.093100 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkm2s\" (UniqueName: \"kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.093133 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.093175 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.093227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.097353 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s" (OuterVolumeSpecName: "kube-api-access-bkm2s") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "kube-api-access-bkm2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.137281 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.137333 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.137304 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.140631 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.146269 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config" (OuterVolumeSpecName: "config") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.196975 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.197005 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkm2s\" (UniqueName: \"kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.197017 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.197027 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.197036 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.197044 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.375418 4824 generic.go:334] "Generic (PLEG): container finished" podID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerID="8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575" exitCode=0 Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.375460 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" event={"ID":"8ad8dd9e-6fe4-462b-98a9-d0192072589d","Type":"ContainerDied","Data":"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575"} Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.375483 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" event={"ID":"8ad8dd9e-6fe4-462b-98a9-d0192072589d","Type":"ContainerDied","Data":"6738b1680c3c03cab4e1e7a4ccdc6941fbdb963c0035872cd1e033904215c51b"} Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.375499 4824 scope.go:117] "RemoveContainer" containerID="8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.375592 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.393511 4824 scope.go:117] "RemoveContainer" containerID="ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.402292 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.409051 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.416111 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-26d8b"] Jan 21 11:30:18 crc kubenswrapper[4824]: W0121 11:30:18.419115 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod976a6446_e79c_42db_be45_43147767f1da.slice/crio-c4cb9cd169639fc7ee58ef770aad0eba3e444be63bdd9e108aa65f4d2ec941f5 WatchSource:0}: Error finding container c4cb9cd169639fc7ee58ef770aad0eba3e444be63bdd9e108aa65f4d2ec941f5: Status 404 returned error can't find the container with id c4cb9cd169639fc7ee58ef770aad0eba3e444be63bdd9e108aa65f4d2ec941f5 Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.426784 4824 scope.go:117] "RemoveContainer" containerID="8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575" Jan 21 11:30:18 crc kubenswrapper[4824]: E0121 11:30:18.427099 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575\": container with ID starting with 8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575 not found: ID does not exist" containerID="8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.427130 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575"} err="failed to get container status \"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575\": rpc error: code = NotFound desc = could not find container \"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575\": container with ID starting with 8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575 not found: ID does not exist" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.427150 4824 scope.go:117] "RemoveContainer" containerID="ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f" Jan 21 11:30:18 crc kubenswrapper[4824]: E0121 11:30:18.427325 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f\": container with ID starting with ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f not found: ID does not exist" containerID="ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.427358 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f"} err="failed to get container status \"ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f\": rpc error: code = NotFound desc = could not find container \"ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f\": container with ID starting with ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f not found: ID does not exist" Jan 21 11:30:19 crc kubenswrapper[4824]: I0121 11:30:19.386202 4824 generic.go:334] "Generic (PLEG): container finished" podID="976a6446-e79c-42db-be45-43147767f1da" containerID="5ea28abe017300018ccb30238cf19011b8c323951c42f861b0e10547b86eb301" exitCode=0 Jan 21 11:30:19 crc kubenswrapper[4824]: I0121 11:30:19.386394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" event={"ID":"976a6446-e79c-42db-be45-43147767f1da","Type":"ContainerDied","Data":"5ea28abe017300018ccb30238cf19011b8c323951c42f861b0e10547b86eb301"} Jan 21 11:30:19 crc kubenswrapper[4824]: I0121 11:30:19.386702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" event={"ID":"976a6446-e79c-42db-be45-43147767f1da","Type":"ContainerStarted","Data":"c4cb9cd169639fc7ee58ef770aad0eba3e444be63bdd9e108aa65f4d2ec941f5"} Jan 21 11:30:20 crc kubenswrapper[4824]: I0121 11:30:20.067783 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" path="/var/lib/kubelet/pods/8ad8dd9e-6fe4-462b-98a9-d0192072589d/volumes" Jan 21 11:30:20 crc kubenswrapper[4824]: I0121 11:30:20.394586 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" event={"ID":"976a6446-e79c-42db-be45-43147767f1da","Type":"ContainerStarted","Data":"8e7c6b4f7a98a2f392bc8913c5a09cf641991cd342a6a59f4cbd9a66fec5cf89"} Jan 21 11:30:20 crc kubenswrapper[4824]: I0121 11:30:20.394985 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:20 crc kubenswrapper[4824]: I0121 11:30:20.414562 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" podStartSLOduration=3.414548463 podStartE2EDuration="3.414548463s" podCreationTimestamp="2026-01-21 11:30:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:30:20.408675849 +0000 UTC m=+1162.701705141" watchObservedRunningTime="2026-01-21 11:30:20.414548463 +0000 UTC m=+1162.707577754" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.003656 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.043150 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.043358 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="dnsmasq-dns" containerID="cri-o://3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f" gracePeriod=10 Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.409058 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.445346 4824 generic.go:334] "Generic (PLEG): container finished" podID="aed87a15-74e4-47e3-8f90-f98672d62792" containerID="3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f" exitCode=0 Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.445386 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.445383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" event={"ID":"aed87a15-74e4-47e3-8f90-f98672d62792","Type":"ContainerDied","Data":"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f"} Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.445657 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" event={"ID":"aed87a15-74e4-47e3-8f90-f98672d62792","Type":"ContainerDied","Data":"d3d7d6a8cb817cd1fffab43e3dc0bcf2917c60a0925857f2eded916d57a51538"} Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.445682 4824 scope.go:117] "RemoveContainer" containerID="3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.464291 4824 scope.go:117] "RemoveContainer" containerID="2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.480332 4824 scope.go:117] "RemoveContainer" containerID="3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f" Jan 21 11:30:28 crc kubenswrapper[4824]: E0121 11:30:28.480619 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f\": container with ID starting with 3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f not found: ID does not exist" containerID="3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.480650 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f"} err="failed to get container status \"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f\": rpc error: code = NotFound desc = could not find container \"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f\": container with ID starting with 3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f not found: ID does not exist" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.480670 4824 scope.go:117] "RemoveContainer" containerID="2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de" Jan 21 11:30:28 crc kubenswrapper[4824]: E0121 11:30:28.480920 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de\": container with ID starting with 2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de not found: ID does not exist" containerID="2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.480947 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de"} err="failed to get container status \"2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de\": rpc error: code = NotFound desc = could not find container \"2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de\": container with ID starting with 2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de not found: ID does not exist" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560649 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560728 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560784 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560843 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560987 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txbf7\" (UniqueName: \"kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.561069 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.564900 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7" (OuterVolumeSpecName: "kube-api-access-txbf7") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "kube-api-access-txbf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.602551 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.615555 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.615566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config" (OuterVolumeSpecName: "config") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.616142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.616158 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.618461 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663521 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txbf7\" (UniqueName: \"kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663549 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663560 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663591 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663600 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663609 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663616 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.769325 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.776265 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:30 crc kubenswrapper[4824]: I0121 11:30:30.057375 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" path="/var/lib/kubelet/pods/aed87a15-74e4-47e3-8f90-f98672d62792/volumes" Jan 21 11:30:39 crc kubenswrapper[4824]: I0121 11:30:39.524967 4824 generic.go:334] "Generic (PLEG): container finished" podID="e4a4d725-1a49-4292-af74-6ab09ab15cdb" containerID="8915f56cc0770d64ee6a0d4ce2d2834b7ba05053d7b8c9baa030f6a50a88dee1" exitCode=0 Jan 21 11:30:39 crc kubenswrapper[4824]: I0121 11:30:39.524986 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4a4d725-1a49-4292-af74-6ab09ab15cdb","Type":"ContainerDied","Data":"8915f56cc0770d64ee6a0d4ce2d2834b7ba05053d7b8c9baa030f6a50a88dee1"} Jan 21 11:30:40 crc kubenswrapper[4824]: I0121 11:30:40.533785 4824 generic.go:334] "Generic (PLEG): container finished" podID="5d8e256a-aa1b-4ccb-8e8c-c86be1e31762" containerID="1ffd78c66a880f9e4de61c175a16fa8e25a34ed2dd520861ea49b68a2f4172fe" exitCode=0 Jan 21 11:30:40 crc kubenswrapper[4824]: I0121 11:30:40.533872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762","Type":"ContainerDied","Data":"1ffd78c66a880f9e4de61c175a16fa8e25a34ed2dd520861ea49b68a2f4172fe"} Jan 21 11:30:40 crc kubenswrapper[4824]: I0121 11:30:40.536832 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4a4d725-1a49-4292-af74-6ab09ab15cdb","Type":"ContainerStarted","Data":"43a3bc01f1993863308a904ac6733e32020664a934b3e525c85fa49c95638ce6"} Jan 21 11:30:40 crc kubenswrapper[4824]: I0121 11:30:40.537005 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 21 11:30:40 crc kubenswrapper[4824]: I0121 11:30:40.576291 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=34.576277187 podStartE2EDuration="34.576277187s" podCreationTimestamp="2026-01-21 11:30:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:30:40.567911366 +0000 UTC m=+1182.860940658" watchObservedRunningTime="2026-01-21 11:30:40.576277187 +0000 UTC m=+1182.869306479" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154100 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf"] Jan 21 11:30:41 crc kubenswrapper[4824]: E0121 11:30:41.154662 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="init" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154681 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="init" Jan 21 11:30:41 crc kubenswrapper[4824]: E0121 11:30:41.154696 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154702 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: E0121 11:30:41.154714 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="init" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154719 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="init" Jan 21 11:30:41 crc kubenswrapper[4824]: E0121 11:30:41.154749 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154754 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154930 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154950 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.155488 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.156843 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.157396 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.160795 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.160836 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.160792 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf"] Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.267033 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hwm7\" (UniqueName: \"kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.267259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.267422 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.267500 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.368740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.368933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.369207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hwm7\" (UniqueName: \"kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.369305 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.372614 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.373216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.373490 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.388764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hwm7\" (UniqueName: \"kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.470463 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.571285 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762","Type":"ContainerStarted","Data":"bb6a9a152262299cc82912966110b0d79c9feaa705bae5ac5405357e09be0b2c"} Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.571578 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.589267 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.589252015 podStartE2EDuration="35.589252015s" podCreationTimestamp="2026-01-21 11:30:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:30:41.589046087 +0000 UTC m=+1183.882075379" watchObservedRunningTime="2026-01-21 11:30:41.589252015 +0000 UTC m=+1183.882281307" Jan 21 11:30:41 crc kubenswrapper[4824]: W0121 11:30:41.908095 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3664d03_d53d_4a49_8cab_c4266f2d1426.slice/crio-a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960 WatchSource:0}: Error finding container a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960: Status 404 returned error can't find the container with id a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960 Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.909586 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.910333 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf"] Jan 21 11:30:42 crc kubenswrapper[4824]: I0121 11:30:42.578213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" event={"ID":"e3664d03-d53d-4a49-8cab-c4266f2d1426","Type":"ContainerStarted","Data":"a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960"} Jan 21 11:30:49 crc kubenswrapper[4824]: I0121 11:30:49.628798 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" event={"ID":"e3664d03-d53d-4a49-8cab-c4266f2d1426","Type":"ContainerStarted","Data":"26eb66e7a517666b02ef2ea7752b23c05bcd7e5afdc97bb9b0069008f7ff7fc5"} Jan 21 11:30:49 crc kubenswrapper[4824]: I0121 11:30:49.643265 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" podStartSLOduration=1.337568133 podStartE2EDuration="8.64325259s" podCreationTimestamp="2026-01-21 11:30:41 +0000 UTC" firstStartedPulling="2026-01-21 11:30:41.909384791 +0000 UTC m=+1184.202414083" lastFinishedPulling="2026-01-21 11:30:49.215069258 +0000 UTC m=+1191.508098540" observedRunningTime="2026-01-21 11:30:49.638516724 +0000 UTC m=+1191.931546015" watchObservedRunningTime="2026-01-21 11:30:49.64325259 +0000 UTC m=+1191.936281883" Jan 21 11:30:56 crc kubenswrapper[4824]: I0121 11:30:56.618093 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 21 11:30:57 crc kubenswrapper[4824]: I0121 11:30:57.014144 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:31:00 crc kubenswrapper[4824]: I0121 11:31:00.708514 4824 generic.go:334] "Generic (PLEG): container finished" podID="e3664d03-d53d-4a49-8cab-c4266f2d1426" containerID="26eb66e7a517666b02ef2ea7752b23c05bcd7e5afdc97bb9b0069008f7ff7fc5" exitCode=0 Jan 21 11:31:00 crc kubenswrapper[4824]: I0121 11:31:00.708594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" event={"ID":"e3664d03-d53d-4a49-8cab-c4266f2d1426","Type":"ContainerDied","Data":"26eb66e7a517666b02ef2ea7752b23c05bcd7e5afdc97bb9b0069008f7ff7fc5"} Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.018593 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.161847 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hwm7\" (UniqueName: \"kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7\") pod \"e3664d03-d53d-4a49-8cab-c4266f2d1426\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.161932 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle\") pod \"e3664d03-d53d-4a49-8cab-c4266f2d1426\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.162039 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam\") pod \"e3664d03-d53d-4a49-8cab-c4266f2d1426\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.162076 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory\") pod \"e3664d03-d53d-4a49-8cab-c4266f2d1426\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.166405 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7" (OuterVolumeSpecName: "kube-api-access-4hwm7") pod "e3664d03-d53d-4a49-8cab-c4266f2d1426" (UID: "e3664d03-d53d-4a49-8cab-c4266f2d1426"). InnerVolumeSpecName "kube-api-access-4hwm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.166692 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e3664d03-d53d-4a49-8cab-c4266f2d1426" (UID: "e3664d03-d53d-4a49-8cab-c4266f2d1426"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.183276 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e3664d03-d53d-4a49-8cab-c4266f2d1426" (UID: "e3664d03-d53d-4a49-8cab-c4266f2d1426"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.183904 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory" (OuterVolumeSpecName: "inventory") pod "e3664d03-d53d-4a49-8cab-c4266f2d1426" (UID: "e3664d03-d53d-4a49-8cab-c4266f2d1426"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.264462 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hwm7\" (UniqueName: \"kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.264646 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.264657 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.264685 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.723162 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" event={"ID":"e3664d03-d53d-4a49-8cab-c4266f2d1426","Type":"ContainerDied","Data":"a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960"} Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.723207 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.723287 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.774923 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn"] Jan 21 11:31:02 crc kubenswrapper[4824]: E0121 11:31:02.775237 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3664d03-d53d-4a49-8cab-c4266f2d1426" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.775254 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3664d03-d53d-4a49-8cab-c4266f2d1426" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.775417 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3664d03-d53d-4a49-8cab-c4266f2d1426" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.775879 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.785111 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.785233 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.785305 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.785434 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.793504 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn"] Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.973905 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.974041 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.974266 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n7ls\" (UniqueName: \"kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.075217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n7ls\" (UniqueName: \"kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.075901 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.076003 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.079586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.079602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.088457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n7ls\" (UniqueName: \"kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.096376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.523746 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn"] Jan 21 11:31:03 crc kubenswrapper[4824]: W0121 11:31:03.525472 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34c68ecf_cf29_44cb_9880_a3d45f625454.slice/crio-e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc WatchSource:0}: Error finding container e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc: Status 404 returned error can't find the container with id e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.730103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" event={"ID":"34c68ecf-cf29-44cb-9880-a3d45f625454","Type":"ContainerStarted","Data":"e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc"} Jan 21 11:31:04 crc kubenswrapper[4824]: I0121 11:31:04.737700 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" event={"ID":"34c68ecf-cf29-44cb-9880-a3d45f625454","Type":"ContainerStarted","Data":"2a39f90f288e86ab7f527b5bc67bfc6b30d16751b1be1f809ef23d416cfc707c"} Jan 21 11:31:04 crc kubenswrapper[4824]: I0121 11:31:04.752320 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" podStartSLOduration=2.231292759 podStartE2EDuration="2.752305525s" podCreationTimestamp="2026-01-21 11:31:02 +0000 UTC" firstStartedPulling="2026-01-21 11:31:03.527609308 +0000 UTC m=+1205.820638600" lastFinishedPulling="2026-01-21 11:31:04.048622075 +0000 UTC m=+1206.341651366" observedRunningTime="2026-01-21 11:31:04.748108754 +0000 UTC m=+1207.041138046" watchObservedRunningTime="2026-01-21 11:31:04.752305525 +0000 UTC m=+1207.045334817" Jan 21 11:31:06 crc kubenswrapper[4824]: I0121 11:31:06.752212 4824 generic.go:334] "Generic (PLEG): container finished" podID="34c68ecf-cf29-44cb-9880-a3d45f625454" containerID="2a39f90f288e86ab7f527b5bc67bfc6b30d16751b1be1f809ef23d416cfc707c" exitCode=0 Jan 21 11:31:06 crc kubenswrapper[4824]: I0121 11:31:06.752302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" event={"ID":"34c68ecf-cf29-44cb-9880-a3d45f625454","Type":"ContainerDied","Data":"2a39f90f288e86ab7f527b5bc67bfc6b30d16751b1be1f809ef23d416cfc707c"} Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.063258 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.257276 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory\") pod \"34c68ecf-cf29-44cb-9880-a3d45f625454\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.257433 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam\") pod \"34c68ecf-cf29-44cb-9880-a3d45f625454\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.257503 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n7ls\" (UniqueName: \"kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls\") pod \"34c68ecf-cf29-44cb-9880-a3d45f625454\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.261552 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls" (OuterVolumeSpecName: "kube-api-access-2n7ls") pod "34c68ecf-cf29-44cb-9880-a3d45f625454" (UID: "34c68ecf-cf29-44cb-9880-a3d45f625454"). InnerVolumeSpecName "kube-api-access-2n7ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.277424 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "34c68ecf-cf29-44cb-9880-a3d45f625454" (UID: "34c68ecf-cf29-44cb-9880-a3d45f625454"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.280832 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory" (OuterVolumeSpecName: "inventory") pod "34c68ecf-cf29-44cb-9880-a3d45f625454" (UID: "34c68ecf-cf29-44cb-9880-a3d45f625454"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.359619 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.359649 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n7ls\" (UniqueName: \"kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.359660 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.768542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" event={"ID":"34c68ecf-cf29-44cb-9880-a3d45f625454","Type":"ContainerDied","Data":"e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc"} Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.768745 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.768596 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.808417 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6"] Jan 21 11:31:08 crc kubenswrapper[4824]: E0121 11:31:08.808788 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c68ecf-cf29-44cb-9880-a3d45f625454" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.808806 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c68ecf-cf29-44cb-9880-a3d45f625454" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.809011 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c68ecf-cf29-44cb-9880-a3d45f625454" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.809605 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.815614 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.815669 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.815827 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.815852 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.816182 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6"] Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.866665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.866854 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.867021 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srdjg\" (UniqueName: \"kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.867063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.969532 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srdjg\" (UniqueName: \"kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.969594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.969731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.969889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.974025 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.979756 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.982876 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.987061 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srdjg\" (UniqueName: \"kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:09 crc kubenswrapper[4824]: I0121 11:31:09.123908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:09 crc kubenswrapper[4824]: I0121 11:31:09.619332 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6"] Jan 21 11:31:09 crc kubenswrapper[4824]: I0121 11:31:09.776695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" event={"ID":"300a037e-ac25-4a9e-9339-6cc707969f7c","Type":"ContainerStarted","Data":"3606afc5a84332a98e06b75db8f8ca12eba78a06f6dfd19a1acd5dec3cb653f4"} Jan 21 11:31:10 crc kubenswrapper[4824]: I0121 11:31:10.783922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" event={"ID":"300a037e-ac25-4a9e-9339-6cc707969f7c","Type":"ContainerStarted","Data":"36e8b1c82d028e5d8507608d6152420eb1389504eb7eb596bd23d52fdf2184e2"} Jan 21 11:31:10 crc kubenswrapper[4824]: I0121 11:31:10.796287 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" podStartSLOduration=2.170215529 podStartE2EDuration="2.796273247s" podCreationTimestamp="2026-01-21 11:31:08 +0000 UTC" firstStartedPulling="2026-01-21 11:31:09.621684068 +0000 UTC m=+1211.914713360" lastFinishedPulling="2026-01-21 11:31:10.247741786 +0000 UTC m=+1212.540771078" observedRunningTime="2026-01-21 11:31:10.793689277 +0000 UTC m=+1213.086718589" watchObservedRunningTime="2026-01-21 11:31:10.796273247 +0000 UTC m=+1213.089302539" Jan 21 11:31:58 crc kubenswrapper[4824]: I0121 11:31:58.939313 4824 scope.go:117] "RemoveContainer" containerID="26ba4c23450a1625fb62c54cdc290126e5f1371f38bf091a20d00cb1c946e58f" Jan 21 11:32:16 crc kubenswrapper[4824]: I0121 11:32:16.065104 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:32:16 crc kubenswrapper[4824]: I0121 11:32:16.065629 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:32:46 crc kubenswrapper[4824]: I0121 11:32:46.064777 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:32:46 crc kubenswrapper[4824]: I0121 11:32:46.065308 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:32:58 crc kubenswrapper[4824]: I0121 11:32:58.985276 4824 scope.go:117] "RemoveContainer" containerID="7802d3564cb8577bdb2e0cb7acab4d7e0721818cb3c2ece34e851778aff1034e" Jan 21 11:32:59 crc kubenswrapper[4824]: I0121 11:32:59.004242 4824 scope.go:117] "RemoveContainer" containerID="eb6d7862490b4afbdbbe422c8984287dc94d6e4a39261b3846440fb920c542fb" Jan 21 11:32:59 crc kubenswrapper[4824]: I0121 11:32:59.036218 4824 scope.go:117] "RemoveContainer" containerID="68285af994e4fe82e817010ea6bff330d5655f091788a58ca5b521d0e79564f1" Jan 21 11:32:59 crc kubenswrapper[4824]: I0121 11:32:59.067153 4824 scope.go:117] "RemoveContainer" containerID="ff804ebe93733fa9e697132922f8e30c6ee75d058b7cb8afecec7da791271c35" Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.064969 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.065476 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.065514 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.066048 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.066100 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4" gracePeriod=600 Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.616564 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4" exitCode=0 Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.616616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4"} Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.617035 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499"} Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.617059 4824 scope.go:117] "RemoveContainer" containerID="74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690" Jan 21 11:34:08 crc kubenswrapper[4824]: I0121 11:34:08.971176 4824 generic.go:334] "Generic (PLEG): container finished" podID="300a037e-ac25-4a9e-9339-6cc707969f7c" containerID="36e8b1c82d028e5d8507608d6152420eb1389504eb7eb596bd23d52fdf2184e2" exitCode=0 Jan 21 11:34:08 crc kubenswrapper[4824]: I0121 11:34:08.971248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" event={"ID":"300a037e-ac25-4a9e-9339-6cc707969f7c","Type":"ContainerDied","Data":"36e8b1c82d028e5d8507608d6152420eb1389504eb7eb596bd23d52fdf2184e2"} Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.278164 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.303178 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory\") pod \"300a037e-ac25-4a9e-9339-6cc707969f7c\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.303300 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam\") pod \"300a037e-ac25-4a9e-9339-6cc707969f7c\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.303360 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle\") pod \"300a037e-ac25-4a9e-9339-6cc707969f7c\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.303471 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srdjg\" (UniqueName: \"kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg\") pod \"300a037e-ac25-4a9e-9339-6cc707969f7c\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.308769 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "300a037e-ac25-4a9e-9339-6cc707969f7c" (UID: "300a037e-ac25-4a9e-9339-6cc707969f7c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.308864 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg" (OuterVolumeSpecName: "kube-api-access-srdjg") pod "300a037e-ac25-4a9e-9339-6cc707969f7c" (UID: "300a037e-ac25-4a9e-9339-6cc707969f7c"). InnerVolumeSpecName "kube-api-access-srdjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.325497 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory" (OuterVolumeSpecName: "inventory") pod "300a037e-ac25-4a9e-9339-6cc707969f7c" (UID: "300a037e-ac25-4a9e-9339-6cc707969f7c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.327398 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "300a037e-ac25-4a9e-9339-6cc707969f7c" (UID: "300a037e-ac25-4a9e-9339-6cc707969f7c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.405519 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.405641 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.405700 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.405766 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srdjg\" (UniqueName: \"kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg\") on node \"crc\" DevicePath \"\"" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.986479 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" event={"ID":"300a037e-ac25-4a9e-9339-6cc707969f7c","Type":"ContainerDied","Data":"3606afc5a84332a98e06b75db8f8ca12eba78a06f6dfd19a1acd5dec3cb653f4"} Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.986867 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3606afc5a84332a98e06b75db8f8ca12eba78a06f6dfd19a1acd5dec3cb653f4" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.986524 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.047720 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9"] Jan 21 11:34:11 crc kubenswrapper[4824]: E0121 11:34:11.048129 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300a037e-ac25-4a9e-9339-6cc707969f7c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.048149 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="300a037e-ac25-4a9e-9339-6cc707969f7c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.048337 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="300a037e-ac25-4a9e-9339-6cc707969f7c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.049075 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.051351 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.051481 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.051605 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.052370 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.055545 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9"] Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.119716 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.120393 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5hlg\" (UniqueName: \"kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.120483 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.222491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.222696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5hlg\" (UniqueName: \"kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.222765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.225662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.225771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.235755 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5hlg\" (UniqueName: \"kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.381369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.797190 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9"] Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.993368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" event={"ID":"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b","Type":"ContainerStarted","Data":"b5d574baa9ddaffa59e308d4f888821c6b55c1c17bdef6ce2814d173fcd2ffdc"} Jan 21 11:34:13 crc kubenswrapper[4824]: I0121 11:34:13.003746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" event={"ID":"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b","Type":"ContainerStarted","Data":"4d04c081499dd9f0ea8ea5831aa73fabb1773d706814fc10d98f8f23c9450fd3"} Jan 21 11:34:13 crc kubenswrapper[4824]: I0121 11:34:13.022512 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" podStartSLOduration=1.552950284 podStartE2EDuration="2.022497125s" podCreationTimestamp="2026-01-21 11:34:11 +0000 UTC" firstStartedPulling="2026-01-21 11:34:11.803496443 +0000 UTC m=+1394.096525736" lastFinishedPulling="2026-01-21 11:34:12.273043285 +0000 UTC m=+1394.566072577" observedRunningTime="2026-01-21 11:34:13.019043938 +0000 UTC m=+1395.312073229" watchObservedRunningTime="2026-01-21 11:34:13.022497125 +0000 UTC m=+1395.315526417" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.141123 4824 scope.go:117] "RemoveContainer" containerID="3f015c1fd2be4ed1835afc4169552ab51d97958ba2fb966dc794c0c63861420b" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.160375 4824 scope.go:117] "RemoveContainer" containerID="87edd33f33351d88d9cbd6b17c925a8893b337952321dc91c1ba77e5c174ae3c" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.192276 4824 scope.go:117] "RemoveContainer" containerID="d91972673e1ac4ea40f718ccf4ab2ec2575e9319007c3a3cc55cff5c9b8fc08a" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.209011 4824 scope.go:117] "RemoveContainer" containerID="97c437ef91be90c480bbbc66844b2f1bf63fc92b3d5efc241a8ff7639e0f322c" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.225644 4824 scope.go:117] "RemoveContainer" containerID="82b46e3d25bb50cb39acd7bbf34eb02ad735f71fcba98f606eba2559159f21c7" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.241507 4824 scope.go:117] "RemoveContainer" containerID="32fb68e8168ce229b5eee2857a8b6f19fa0565f1c232d572bde25ae83e70fb37" Jan 21 11:35:16 crc kubenswrapper[4824]: I0121 11:35:16.064714 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:35:16 crc kubenswrapper[4824]: I0121 11:35:16.065456 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:35:35 crc kubenswrapper[4824]: I0121 11:35:35.559761 4824 generic.go:334] "Generic (PLEG): container finished" podID="8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" containerID="4d04c081499dd9f0ea8ea5831aa73fabb1773d706814fc10d98f8f23c9450fd3" exitCode=0 Jan 21 11:35:35 crc kubenswrapper[4824]: I0121 11:35:35.559832 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" event={"ID":"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b","Type":"ContainerDied","Data":"4d04c081499dd9f0ea8ea5831aa73fabb1773d706814fc10d98f8f23c9450fd3"} Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.868476 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.956249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam\") pod \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.956335 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory\") pod \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.956411 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5hlg\" (UniqueName: \"kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg\") pod \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.961463 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg" (OuterVolumeSpecName: "kube-api-access-v5hlg") pod "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" (UID: "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b"). InnerVolumeSpecName "kube-api-access-v5hlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.976976 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory" (OuterVolumeSpecName: "inventory") pod "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" (UID: "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.978441 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" (UID: "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.059071 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.059101 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5hlg\" (UniqueName: \"kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg\") on node \"crc\" DevicePath \"\"" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.059113 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.574341 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" event={"ID":"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b","Type":"ContainerDied","Data":"b5d574baa9ddaffa59e308d4f888821c6b55c1c17bdef6ce2814d173fcd2ffdc"} Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.574578 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5d574baa9ddaffa59e308d4f888821c6b55c1c17bdef6ce2814d173fcd2ffdc" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.574405 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.634051 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz"] Jan 21 11:35:37 crc kubenswrapper[4824]: E0121 11:35:37.634415 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.634433 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.634652 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.635257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.636588 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.637184 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.637518 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.637544 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.644723 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz"] Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.765621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.765750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.765874 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m499z\" (UniqueName: \"kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.868114 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.868206 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m499z\" (UniqueName: \"kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.868249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.871690 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.871761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.881347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m499z\" (UniqueName: \"kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.954311 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.043595 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jbk7q"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.061857 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a0fd-account-create-update-zxjvg"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.061884 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-q7wdg"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.065430 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-383b-account-create-update-w99bv"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.075039 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a0fd-account-create-update-zxjvg"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.082220 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-383b-account-create-update-w99bv"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.089529 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-q7wdg"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.095850 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jbk7q"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.388973 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz"] Jan 21 11:35:38 crc kubenswrapper[4824]: W0121 11:35:38.396022 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8c7ec7b_ee3f_46bc_95f8_4406713ba43a.slice/crio-64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289 WatchSource:0}: Error finding container 64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289: Status 404 returned error can't find the container with id 64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289 Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.581337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" event={"ID":"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a","Type":"ContainerStarted","Data":"64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289"} Jan 21 11:35:39 crc kubenswrapper[4824]: I0121 11:35:39.588469 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" event={"ID":"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a","Type":"ContainerStarted","Data":"49f4bbc73004e2aa4fd9edf12e2c015dacbc34505e45f5cb367148bf340ae9dd"} Jan 21 11:35:39 crc kubenswrapper[4824]: I0121 11:35:39.600598 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" podStartSLOduration=2.101713229 podStartE2EDuration="2.600585481s" podCreationTimestamp="2026-01-21 11:35:37 +0000 UTC" firstStartedPulling="2026-01-21 11:35:38.397522041 +0000 UTC m=+1480.690551333" lastFinishedPulling="2026-01-21 11:35:38.896394292 +0000 UTC m=+1481.189423585" observedRunningTime="2026-01-21 11:35:39.597452125 +0000 UTC m=+1481.890481417" watchObservedRunningTime="2026-01-21 11:35:39.600585481 +0000 UTC m=+1481.893614773" Jan 21 11:35:40 crc kubenswrapper[4824]: I0121 11:35:40.057651 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c01cf79-8a28-4229-9257-90124e66d3cc" path="/var/lib/kubelet/pods/7c01cf79-8a28-4229-9257-90124e66d3cc/volumes" Jan 21 11:35:40 crc kubenswrapper[4824]: I0121 11:35:40.058294 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="885db96c-779f-47c5-a383-1821cb53977c" path="/var/lib/kubelet/pods/885db96c-779f-47c5-a383-1821cb53977c/volumes" Jan 21 11:35:40 crc kubenswrapper[4824]: I0121 11:35:40.058796 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" path="/var/lib/kubelet/pods/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01/volumes" Jan 21 11:35:40 crc kubenswrapper[4824]: I0121 11:35:40.059334 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc853ab0-e92a-4ada-84fd-b59ac2af573d" path="/var/lib/kubelet/pods/bc853ab0-e92a-4ada-84fd-b59ac2af573d/volumes" Jan 21 11:35:44 crc kubenswrapper[4824]: I0121 11:35:44.022300 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-rlxbb"] Jan 21 11:35:44 crc kubenswrapper[4824]: I0121 11:35:44.029796 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-rlxbb"] Jan 21 11:35:44 crc kubenswrapper[4824]: I0121 11:35:44.037672 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-0697-account-create-update-9nvnh"] Jan 21 11:35:44 crc kubenswrapper[4824]: I0121 11:35:44.068316 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fbc8b12-d965-44e9-822f-b95a5c5e88ab" path="/var/lib/kubelet/pods/6fbc8b12-d965-44e9-822f-b95a5c5e88ab/volumes" Jan 21 11:35:44 crc kubenswrapper[4824]: I0121 11:35:44.069047 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-0697-account-create-update-9nvnh"] Jan 21 11:35:46 crc kubenswrapper[4824]: I0121 11:35:46.056979 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72452c85-3787-4bd5-b605-af5555c1fdaf" path="/var/lib/kubelet/pods/72452c85-3787-4bd5-b605-af5555c1fdaf/volumes" Jan 21 11:35:46 crc kubenswrapper[4824]: I0121 11:35:46.064791 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:35:46 crc kubenswrapper[4824]: I0121 11:35:46.064859 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.045813 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-m6rpv"] Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.052298 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-m6rpv"] Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.288562 4824 scope.go:117] "RemoveContainer" containerID="aece7515cad2f167dd4b5a5e58d5aa5c0df426c12c8bbc743e03dad13feaedf7" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.318906 4824 scope.go:117] "RemoveContainer" containerID="0811ac195531b85d7ce2dbbbbfe5e4b4eec9c6e973dbfd91e9e942aecc0be98b" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.340093 4824 scope.go:117] "RemoveContainer" containerID="091c53575308b0937e0789468c0edd3eb80a63b653591b4cc1559689a8b4b286" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.380058 4824 scope.go:117] "RemoveContainer" containerID="7d54dd5634955be1b99e4810715039653beb466487ffbc563c911444570d3daa" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.432740 4824 scope.go:117] "RemoveContainer" containerID="be822998cfd0d1fe00b165b7bdf2ce2c669d12e3704320cfa573970aa648494d" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.461559 4824 scope.go:117] "RemoveContainer" containerID="e85df67ff5e42c058f160ae767051c544cbf1f8e15561d3884068b5517d68c3a" Jan 21 11:36:00 crc kubenswrapper[4824]: I0121 11:36:00.056758 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76e33763-1a41-4640-8e4a-492ab92009e8" path="/var/lib/kubelet/pods/76e33763-1a41-4640-8e4a-492ab92009e8/volumes" Jan 21 11:36:04 crc kubenswrapper[4824]: I0121 11:36:04.020061 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-sx9lb"] Jan 21 11:36:04 crc kubenswrapper[4824]: I0121 11:36:04.026550 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-sx9lb"] Jan 21 11:36:04 crc kubenswrapper[4824]: I0121 11:36:04.057148 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee7a30d4-dcec-4719-a629-dfaf588d4169" path="/var/lib/kubelet/pods/ee7a30d4-dcec-4719-a629-dfaf588d4169/volumes" Jan 21 11:36:15 crc kubenswrapper[4824]: I0121 11:36:15.025259 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-2619-account-create-update-9csh2"] Jan 21 11:36:15 crc kubenswrapper[4824]: I0121 11:36:15.035175 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-kk8kb"] Jan 21 11:36:15 crc kubenswrapper[4824]: I0121 11:36:15.046603 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-kk8kb"] Jan 21 11:36:15 crc kubenswrapper[4824]: I0121 11:36:15.060730 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-2619-account-create-update-9csh2"] Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.056928 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8600373d-f466-43d4-92fc-1fa938f6e91b" path="/var/lib/kubelet/pods/8600373d-f466-43d4-92fc-1fa938f6e91b/volumes" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.057715 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe6e7565-793b-4555-b6d5-758f87a5b9c7" path="/var/lib/kubelet/pods/fe6e7565-793b-4555-b6d5-758f87a5b9c7/volumes" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.064732 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.064780 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.064816 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.065223 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.065277 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" gracePeriod=600 Jan 21 11:36:16 crc kubenswrapper[4824]: E0121 11:36:16.191313 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.827867 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" exitCode=0 Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.827920 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499"} Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.828133 4824 scope.go:117] "RemoveContainer" containerID="f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.828705 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:36:16 crc kubenswrapper[4824]: E0121 11:36:16.828981 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.096453 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-twgjw"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.105753 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-60f6-account-create-update-pjh6n"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.112131 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-ngtnj"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.118227 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-60f6-account-create-update-pjh6n"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.123986 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-0662-account-create-update-j7bhp"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.128991 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-ngtnj"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.133890 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-twgjw"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.138859 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-3c9f-account-create-update-rtprr"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.143853 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-0662-account-create-update-j7bhp"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.148774 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-3c9f-account-create-update-rtprr"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.153687 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-nskxc"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.158610 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-nskxc"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.239456 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.241029 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.247588 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.379741 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.379876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn7n7\" (UniqueName: \"kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.379942 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.481591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.481700 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn7n7\" (UniqueName: \"kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.481764 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.482142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.482360 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.498148 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn7n7\" (UniqueName: \"kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.554724 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:19 crc kubenswrapper[4824]: I0121 11:36:19.072641 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:19 crc kubenswrapper[4824]: I0121 11:36:19.852612 4824 generic.go:334] "Generic (PLEG): container finished" podID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerID="88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d" exitCode=0 Jan 21 11:36:19 crc kubenswrapper[4824]: I0121 11:36:19.852699 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerDied","Data":"88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d"} Jan 21 11:36:19 crc kubenswrapper[4824]: I0121 11:36:19.852825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerStarted","Data":"33b7c4d8b7630213d98e01a69d1c43ab8893a8320c312d1bd10ff4050bcd0312"} Jan 21 11:36:19 crc kubenswrapper[4824]: I0121 11:36:19.854209 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.057453 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c0ddc66-7696-4816-8e96-340aeb39d242" path="/var/lib/kubelet/pods/5c0ddc66-7696-4816-8e96-340aeb39d242/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.058275 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b747df8-e61f-4689-a08c-0dc6688f1a0e" path="/var/lib/kubelet/pods/7b747df8-e61f-4689-a08c-0dc6688f1a0e/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.058815 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" path="/var/lib/kubelet/pods/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.059355 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9be37b35-7ad5-434a-8d16-c2f9f1661821" path="/var/lib/kubelet/pods/9be37b35-7ad5-434a-8d16-c2f9f1661821/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.060297 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a10a8f17-c27b-4544-8f9d-a6afd9991a8e" path="/var/lib/kubelet/pods/a10a8f17-c27b-4544-8f9d-a6afd9991a8e/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.060807 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4773e7a-a50c-42e6-bb27-a25b0055b8e3" path="/var/lib/kubelet/pods/b4773e7a-a50c-42e6-bb27-a25b0055b8e3/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.861359 4824 generic.go:334] "Generic (PLEG): container finished" podID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerID="17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20" exitCode=0 Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.861569 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerDied","Data":"17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20"} Jan 21 11:36:21 crc kubenswrapper[4824]: I0121 11:36:21.030157 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-84f4b"] Jan 21 11:36:21 crc kubenswrapper[4824]: I0121 11:36:21.036435 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-84f4b"] Jan 21 11:36:21 crc kubenswrapper[4824]: I0121 11:36:21.869665 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerStarted","Data":"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829"} Jan 21 11:36:21 crc kubenswrapper[4824]: I0121 11:36:21.887470 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xhcp9" podStartSLOduration=2.4058206540000002 podStartE2EDuration="3.88743833s" podCreationTimestamp="2026-01-21 11:36:18 +0000 UTC" firstStartedPulling="2026-01-21 11:36:19.854000603 +0000 UTC m=+1522.147029895" lastFinishedPulling="2026-01-21 11:36:21.33561828 +0000 UTC m=+1523.628647571" observedRunningTime="2026-01-21 11:36:21.884042188 +0000 UTC m=+1524.177071480" watchObservedRunningTime="2026-01-21 11:36:21.88743833 +0000 UTC m=+1524.180467621" Jan 21 11:36:22 crc kubenswrapper[4824]: I0121 11:36:22.057836 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5240d35b-cbf1-472d-91da-debf418dd208" path="/var/lib/kubelet/pods/5240d35b-cbf1-472d-91da-debf418dd208/volumes" Jan 21 11:36:28 crc kubenswrapper[4824]: I0121 11:36:28.559230 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:28 crc kubenswrapper[4824]: I0121 11:36:28.559631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:28 crc kubenswrapper[4824]: I0121 11:36:28.594355 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:28 crc kubenswrapper[4824]: I0121 11:36:28.955781 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:29 crc kubenswrapper[4824]: I0121 11:36:29.003675 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:29 crc kubenswrapper[4824]: I0121 11:36:29.049295 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:36:29 crc kubenswrapper[4824]: E0121 11:36:29.049644 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:36:30 crc kubenswrapper[4824]: I0121 11:36:30.936296 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xhcp9" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="registry-server" containerID="cri-o://051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829" gracePeriod=2 Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.323840 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.488391 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn7n7\" (UniqueName: \"kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7\") pod \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.488858 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content\") pod \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.489023 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities\") pod \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.489597 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities" (OuterVolumeSpecName: "utilities") pod "edf7c1b7-286d-4325-ade6-e13af1ba3b26" (UID: "edf7c1b7-286d-4325-ade6-e13af1ba3b26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.489791 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.494019 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7" (OuterVolumeSpecName: "kube-api-access-pn7n7") pod "edf7c1b7-286d-4325-ade6-e13af1ba3b26" (UID: "edf7c1b7-286d-4325-ade6-e13af1ba3b26"). InnerVolumeSpecName "kube-api-access-pn7n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.506989 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "edf7c1b7-286d-4325-ade6-e13af1ba3b26" (UID: "edf7c1b7-286d-4325-ade6-e13af1ba3b26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.592119 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.592148 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn7n7\" (UniqueName: \"kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.944632 4824 generic.go:334] "Generic (PLEG): container finished" podID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerID="051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829" exitCode=0 Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.944670 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerDied","Data":"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829"} Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.944693 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerDied","Data":"33b7c4d8b7630213d98e01a69d1c43ab8893a8320c312d1bd10ff4050bcd0312"} Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.944707 4824 scope.go:117] "RemoveContainer" containerID="051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.944802 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.962714 4824 scope.go:117] "RemoveContainer" containerID="17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.976194 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.982635 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.993183 4824 scope.go:117] "RemoveContainer" containerID="88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.016464 4824 scope.go:117] "RemoveContainer" containerID="051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829" Jan 21 11:36:32 crc kubenswrapper[4824]: E0121 11:36:32.019908 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829\": container with ID starting with 051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829 not found: ID does not exist" containerID="051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.019943 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829"} err="failed to get container status \"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829\": rpc error: code = NotFound desc = could not find container \"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829\": container with ID starting with 051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829 not found: ID does not exist" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.020005 4824 scope.go:117] "RemoveContainer" containerID="17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20" Jan 21 11:36:32 crc kubenswrapper[4824]: E0121 11:36:32.020265 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20\": container with ID starting with 17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20 not found: ID does not exist" containerID="17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.020287 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20"} err="failed to get container status \"17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20\": rpc error: code = NotFound desc = could not find container \"17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20\": container with ID starting with 17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20 not found: ID does not exist" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.020299 4824 scope.go:117] "RemoveContainer" containerID="88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d" Jan 21 11:36:32 crc kubenswrapper[4824]: E0121 11:36:32.020892 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d\": container with ID starting with 88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d not found: ID does not exist" containerID="88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.020924 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d"} err="failed to get container status \"88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d\": rpc error: code = NotFound desc = could not find container \"88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d\": container with ID starting with 88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d not found: ID does not exist" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.056724 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" path="/var/lib/kubelet/pods/edf7c1b7-286d-4325-ade6-e13af1ba3b26/volumes" Jan 21 11:36:34 crc kubenswrapper[4824]: I0121 11:36:34.966525 4824 generic.go:334] "Generic (PLEG): container finished" podID="b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" containerID="49f4bbc73004e2aa4fd9edf12e2c015dacbc34505e45f5cb367148bf340ae9dd" exitCode=0 Jan 21 11:36:34 crc kubenswrapper[4824]: I0121 11:36:34.966623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" event={"ID":"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a","Type":"ContainerDied","Data":"49f4bbc73004e2aa4fd9edf12e2c015dacbc34505e45f5cb367148bf340ae9dd"} Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.289841 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.376312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m499z\" (UniqueName: \"kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z\") pod \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.376516 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam\") pod \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.376717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory\") pod \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.382799 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z" (OuterVolumeSpecName: "kube-api-access-m499z") pod "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" (UID: "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a"). InnerVolumeSpecName "kube-api-access-m499z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.400630 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" (UID: "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.406806 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory" (OuterVolumeSpecName: "inventory") pod "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" (UID: "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.479157 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.479198 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.479212 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m499z\" (UniqueName: \"kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.981090 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" event={"ID":"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a","Type":"ContainerDied","Data":"64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289"} Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.981397 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.981165 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044079 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz"] Jan 21 11:36:37 crc kubenswrapper[4824]: E0121 11:36:37.044477 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="extract-utilities" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044498 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="extract-utilities" Jan 21 11:36:37 crc kubenswrapper[4824]: E0121 11:36:37.044507 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="registry-server" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044513 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="registry-server" Jan 21 11:36:37 crc kubenswrapper[4824]: E0121 11:36:37.044532 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044539 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:37 crc kubenswrapper[4824]: E0121 11:36:37.044558 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="extract-content" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044563 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="extract-content" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044757 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="registry-server" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044773 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.045430 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.046861 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.047018 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.047071 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.047541 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.049249 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz"] Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.191383 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92v6n\" (UniqueName: \"kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.191557 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.191618 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.293647 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.293731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.293888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92v6n\" (UniqueName: \"kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.297427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.298865 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.310242 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92v6n\" (UniqueName: \"kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.359744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.803628 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz"] Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.989114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" event={"ID":"d672f153-6ef2-432d-829a-db63c8daab69","Type":"ContainerStarted","Data":"73fe2f366123c47bd83aebfb3c09855e44f3bceec576e057fa2d17a95e4879fb"} Jan 21 11:36:39 crc kubenswrapper[4824]: I0121 11:36:39.001311 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" event={"ID":"d672f153-6ef2-432d-829a-db63c8daab69","Type":"ContainerStarted","Data":"7778ba4059495bd1d394bac5a6c5c17aa054fa3ecf50af111d3388b007e74498"} Jan 21 11:36:39 crc kubenswrapper[4824]: I0121 11:36:39.017068 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" podStartSLOduration=1.483571178 podStartE2EDuration="2.01705351s" podCreationTimestamp="2026-01-21 11:36:37 +0000 UTC" firstStartedPulling="2026-01-21 11:36:37.806452611 +0000 UTC m=+1540.099481892" lastFinishedPulling="2026-01-21 11:36:38.339934931 +0000 UTC m=+1540.632964224" observedRunningTime="2026-01-21 11:36:39.011750658 +0000 UTC m=+1541.304779951" watchObservedRunningTime="2026-01-21 11:36:39.01705351 +0000 UTC m=+1541.310082801" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.031710 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-t2txg"] Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.038574 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-t2txg"] Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.059722 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a19b6730-b7aa-4319-8127-c32d7874a471" path="/var/lib/kubelet/pods/a19b6730-b7aa-4319-8127-c32d7874a471/volumes" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.296453 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.298268 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.312081 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.449260 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.449333 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbmzk\" (UniqueName: \"kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.449603 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.551041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.551201 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.551260 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbmzk\" (UniqueName: \"kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.551538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.551594 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.568655 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbmzk\" (UniqueName: \"kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.613564 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:41 crc kubenswrapper[4824]: I0121 11:36:41.042703 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:42 crc kubenswrapper[4824]: I0121 11:36:42.022318 4824 generic.go:334] "Generic (PLEG): container finished" podID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerID="d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a" exitCode=0 Jan 21 11:36:42 crc kubenswrapper[4824]: I0121 11:36:42.022375 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerDied","Data":"d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a"} Jan 21 11:36:42 crc kubenswrapper[4824]: I0121 11:36:42.022422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerStarted","Data":"a182f73b4f377f419a0b4bf8458079f91f332f55dfb874dca67b3bccb1efe150"} Jan 21 11:36:42 crc kubenswrapper[4824]: I0121 11:36:42.049899 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:36:42 crc kubenswrapper[4824]: E0121 11:36:42.050232 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:36:43 crc kubenswrapper[4824]: I0121 11:36:43.030347 4824 generic.go:334] "Generic (PLEG): container finished" podID="d672f153-6ef2-432d-829a-db63c8daab69" containerID="7778ba4059495bd1d394bac5a6c5c17aa054fa3ecf50af111d3388b007e74498" exitCode=0 Jan 21 11:36:43 crc kubenswrapper[4824]: I0121 11:36:43.030427 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" event={"ID":"d672f153-6ef2-432d-829a-db63c8daab69","Type":"ContainerDied","Data":"7778ba4059495bd1d394bac5a6c5c17aa054fa3ecf50af111d3388b007e74498"} Jan 21 11:36:43 crc kubenswrapper[4824]: I0121 11:36:43.032421 4824 generic.go:334] "Generic (PLEG): container finished" podID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerID="a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1" exitCode=0 Jan 21 11:36:43 crc kubenswrapper[4824]: I0121 11:36:43.032456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerDied","Data":"a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1"} Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.019703 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6vgh6"] Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.025793 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6vgh6"] Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.045464 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerStarted","Data":"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8"} Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.056547 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08b8c363-a8c1-4dfe-8fd1-1af2d055a305" path="/var/lib/kubelet/pods/08b8c363-a8c1-4dfe-8fd1-1af2d055a305/volumes" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.065159 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m5zcg" podStartSLOduration=2.596961728 podStartE2EDuration="4.065146825s" podCreationTimestamp="2026-01-21 11:36:40 +0000 UTC" firstStartedPulling="2026-01-21 11:36:42.024217976 +0000 UTC m=+1544.317247268" lastFinishedPulling="2026-01-21 11:36:43.492403073 +0000 UTC m=+1545.785432365" observedRunningTime="2026-01-21 11:36:44.059852029 +0000 UTC m=+1546.352881321" watchObservedRunningTime="2026-01-21 11:36:44.065146825 +0000 UTC m=+1546.358176116" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.394773 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.518477 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory\") pod \"d672f153-6ef2-432d-829a-db63c8daab69\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.518840 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam\") pod \"d672f153-6ef2-432d-829a-db63c8daab69\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.519201 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92v6n\" (UniqueName: \"kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n\") pod \"d672f153-6ef2-432d-829a-db63c8daab69\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.523054 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n" (OuterVolumeSpecName: "kube-api-access-92v6n") pod "d672f153-6ef2-432d-829a-db63c8daab69" (UID: "d672f153-6ef2-432d-829a-db63c8daab69"). InnerVolumeSpecName "kube-api-access-92v6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.539172 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory" (OuterVolumeSpecName: "inventory") pod "d672f153-6ef2-432d-829a-db63c8daab69" (UID: "d672f153-6ef2-432d-829a-db63c8daab69"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.540941 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d672f153-6ef2-432d-829a-db63c8daab69" (UID: "d672f153-6ef2-432d-829a-db63c8daab69"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.621611 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.621646 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92v6n\" (UniqueName: \"kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.621656 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.053821 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" event={"ID":"d672f153-6ef2-432d-829a-db63c8daab69","Type":"ContainerDied","Data":"73fe2f366123c47bd83aebfb3c09855e44f3bceec576e057fa2d17a95e4879fb"} Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.053863 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73fe2f366123c47bd83aebfb3c09855e44f3bceec576e057fa2d17a95e4879fb" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.054029 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.106650 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh"] Jan 21 11:36:45 crc kubenswrapper[4824]: E0121 11:36:45.107021 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d672f153-6ef2-432d-829a-db63c8daab69" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.107040 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d672f153-6ef2-432d-829a-db63c8daab69" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.107222 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d672f153-6ef2-432d-829a-db63c8daab69" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.107721 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.109873 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.110072 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.110350 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.110498 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.114342 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh"] Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.231615 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h2jt\" (UniqueName: \"kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.231914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.232267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.335170 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.335587 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.336179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h2jt\" (UniqueName: \"kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.338539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.338734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.350401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h2jt\" (UniqueName: \"kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.420590 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.851751 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh"] Jan 21 11:36:45 crc kubenswrapper[4824]: W0121 11:36:45.852472 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60fedbf7_07e6_41be_ada3_2a06f28835e3.slice/crio-bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490 WatchSource:0}: Error finding container bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490: Status 404 returned error can't find the container with id bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490 Jan 21 11:36:46 crc kubenswrapper[4824]: I0121 11:36:46.026314 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5l2gg"] Jan 21 11:36:46 crc kubenswrapper[4824]: I0121 11:36:46.032560 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5l2gg"] Jan 21 11:36:46 crc kubenswrapper[4824]: I0121 11:36:46.057018 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e75011-d633-4ccc-951c-d019cb2100f9" path="/var/lib/kubelet/pods/a7e75011-d633-4ccc-951c-d019cb2100f9/volumes" Jan 21 11:36:46 crc kubenswrapper[4824]: I0121 11:36:46.061549 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" event={"ID":"60fedbf7-07e6-41be-ada3-2a06f28835e3","Type":"ContainerStarted","Data":"bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490"} Jan 21 11:36:47 crc kubenswrapper[4824]: I0121 11:36:47.070670 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" event={"ID":"60fedbf7-07e6-41be-ada3-2a06f28835e3","Type":"ContainerStarted","Data":"495175a3ec4152174c9a4e2a8aae14a46ac8d180433f99dc41b0fab262a48f7c"} Jan 21 11:36:47 crc kubenswrapper[4824]: I0121 11:36:47.086885 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" podStartSLOduration=1.540283025 podStartE2EDuration="2.08687221s" podCreationTimestamp="2026-01-21 11:36:45 +0000 UTC" firstStartedPulling="2026-01-21 11:36:45.854045762 +0000 UTC m=+1548.147075054" lastFinishedPulling="2026-01-21 11:36:46.400634947 +0000 UTC m=+1548.693664239" observedRunningTime="2026-01-21 11:36:47.081806126 +0000 UTC m=+1549.374835418" watchObservedRunningTime="2026-01-21 11:36:47.08687221 +0000 UTC m=+1549.379901502" Jan 21 11:36:48 crc kubenswrapper[4824]: I0121 11:36:48.019345 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-jrsn8"] Jan 21 11:36:48 crc kubenswrapper[4824]: I0121 11:36:48.025309 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-jrsn8"] Jan 21 11:36:48 crc kubenswrapper[4824]: I0121 11:36:48.056764 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cadf2d38-24f9-43ab-8e29-803d97a0ff06" path="/var/lib/kubelet/pods/cadf2d38-24f9-43ab-8e29-803d97a0ff06/volumes" Jan 21 11:36:50 crc kubenswrapper[4824]: I0121 11:36:50.613813 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:50 crc kubenswrapper[4824]: I0121 11:36:50.614070 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:50 crc kubenswrapper[4824]: I0121 11:36:50.644749 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:51 crc kubenswrapper[4824]: I0121 11:36:51.126235 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:51 crc kubenswrapper[4824]: I0121 11:36:51.165228 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.106344 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m5zcg" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="registry-server" containerID="cri-o://10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8" gracePeriod=2 Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.466308 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.668712 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content\") pod \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.669150 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbmzk\" (UniqueName: \"kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk\") pod \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.669218 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities\") pod \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.669646 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities" (OuterVolumeSpecName: "utilities") pod "6ba09de8-56d0-4a49-90fd-0901ef2a4983" (UID: "6ba09de8-56d0-4a49-90fd-0901ef2a4983"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.673727 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk" (OuterVolumeSpecName: "kube-api-access-wbmzk") pod "6ba09de8-56d0-4a49-90fd-0901ef2a4983" (UID: "6ba09de8-56d0-4a49-90fd-0901ef2a4983"). InnerVolumeSpecName "kube-api-access-wbmzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.707884 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ba09de8-56d0-4a49-90fd-0901ef2a4983" (UID: "6ba09de8-56d0-4a49-90fd-0901ef2a4983"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.770750 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbmzk\" (UniqueName: \"kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.770775 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.770785 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.114562 4824 generic.go:334] "Generic (PLEG): container finished" podID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerID="10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8" exitCode=0 Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.114610 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerDied","Data":"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8"} Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.114641 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerDied","Data":"a182f73b4f377f419a0b4bf8458079f91f332f55dfb874dca67b3bccb1efe150"} Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.114658 4824 scope.go:117] "RemoveContainer" containerID="10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.115497 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.130942 4824 scope.go:117] "RemoveContainer" containerID="a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.133289 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.140434 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.144794 4824 scope.go:117] "RemoveContainer" containerID="d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.176018 4824 scope.go:117] "RemoveContainer" containerID="10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8" Jan 21 11:36:54 crc kubenswrapper[4824]: E0121 11:36:54.176329 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8\": container with ID starting with 10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8 not found: ID does not exist" containerID="10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.176362 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8"} err="failed to get container status \"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8\": rpc error: code = NotFound desc = could not find container \"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8\": container with ID starting with 10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8 not found: ID does not exist" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.176383 4824 scope.go:117] "RemoveContainer" containerID="a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1" Jan 21 11:36:54 crc kubenswrapper[4824]: E0121 11:36:54.176688 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1\": container with ID starting with a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1 not found: ID does not exist" containerID="a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.176710 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1"} err="failed to get container status \"a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1\": rpc error: code = NotFound desc = could not find container \"a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1\": container with ID starting with a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1 not found: ID does not exist" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.176725 4824 scope.go:117] "RemoveContainer" containerID="d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a" Jan 21 11:36:54 crc kubenswrapper[4824]: E0121 11:36:54.177218 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a\": container with ID starting with d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a not found: ID does not exist" containerID="d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.177257 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a"} err="failed to get container status \"d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a\": rpc error: code = NotFound desc = could not find container \"d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a\": container with ID starting with d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a not found: ID does not exist" Jan 21 11:36:55 crc kubenswrapper[4824]: I0121 11:36:55.049361 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:36:55 crc kubenswrapper[4824]: E0121 11:36:55.049582 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.057570 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" path="/var/lib/kubelet/pods/6ba09de8-56d0-4a49-90fd-0901ef2a4983/volumes" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.479929 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:36:56 crc kubenswrapper[4824]: E0121 11:36:56.480329 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="extract-utilities" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.480348 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="extract-utilities" Jan 21 11:36:56 crc kubenswrapper[4824]: E0121 11:36:56.480362 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="registry-server" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.480368 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="registry-server" Jan 21 11:36:56 crc kubenswrapper[4824]: E0121 11:36:56.480382 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="extract-content" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.480388 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="extract-content" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.480562 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="registry-server" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.481770 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.487581 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.515782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.515816 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.515875 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rqjc\" (UniqueName: \"kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.617349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.617381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.617437 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rqjc\" (UniqueName: \"kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.617795 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.617816 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.632361 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rqjc\" (UniqueName: \"kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.807850 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:57 crc kubenswrapper[4824]: I0121 11:36:57.224675 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:36:58 crc kubenswrapper[4824]: I0121 11:36:58.142104 4824 generic.go:334] "Generic (PLEG): container finished" podID="5633f007-e1df-4760-b397-db8d77a2c43e" containerID="e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d" exitCode=0 Jan 21 11:36:58 crc kubenswrapper[4824]: I0121 11:36:58.142169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerDied","Data":"e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d"} Jan 21 11:36:58 crc kubenswrapper[4824]: I0121 11:36:58.142321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerStarted","Data":"9730f16fe571b01572a5f22435b85174e662bfa2a84a0568a751de97d620c25c"} Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.596399 4824 scope.go:117] "RemoveContainer" containerID="2176bc084f2009e4d942790b8c3474f286b639f648b347b28da26c0a5a3ccec2" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.628053 4824 scope.go:117] "RemoveContainer" containerID="49abf3af76b6411b5212e8cfa13e8c4145a6700e1d310d3281c5d847c1a1f457" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.663006 4824 scope.go:117] "RemoveContainer" containerID="81f52e55093614da3aed12f82750111917b47ee8401eb58ccc0da58488e3019f" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.683373 4824 scope.go:117] "RemoveContainer" containerID="45d2f067a75588f2f74c579f26d8f9af2b6a872de4eb9d340b34e9fee4726547" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.716479 4824 scope.go:117] "RemoveContainer" containerID="e12a5e35428ef122a88a06ec7747afe8e712c9a7807b731859f6ec595579f719" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.751702 4824 scope.go:117] "RemoveContainer" containerID="2534da156868c09c8dd8dee0424c34fab5c7b1401695a79ef623bde9f1ef00f0" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.781722 4824 scope.go:117] "RemoveContainer" containerID="3d07104d28aceb5ad31acec840c18d559f5c98c4390845b748361ff21f72d204" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.808032 4824 scope.go:117] "RemoveContainer" containerID="cc788777a506818eef849722d9795fcf44c54eb36cb8cfbc69470609fce80bd0" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.822147 4824 scope.go:117] "RemoveContainer" containerID="5dcf0246f4ab2f2f4f3230d6a47ec6ac1102fcb34b4edb4d89a73c99d792f5ff" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.846032 4824 scope.go:117] "RemoveContainer" containerID="5f4a1e6bcd0c138e59cd1e69d47cce9046bae28ab6b9ce57bd7be2ad0c6a4e61" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.865161 4824 scope.go:117] "RemoveContainer" containerID="c1a92fb4372889a2ab5383adea30e43a1fa74caae3039082b2ef5482212d2758" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.880623 4824 scope.go:117] "RemoveContainer" containerID="b6f52bcc83bf92c2ef07afe3fa11fe445d957e311fc35ba4f1767a43394658f9" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.895429 4824 scope.go:117] "RemoveContainer" containerID="f658b544252ce5134cf2b3d03b008e5a8cf0c0a33b02b6a5554a5288bb79bbfe" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.909618 4824 scope.go:117] "RemoveContainer" containerID="25e2c38241236d848b515e556d4747f5cb7cbca9e5eba1926722e4f2e38b424d" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.922127 4824 scope.go:117] "RemoveContainer" containerID="2b4aa0578cb659bd1fe3cb0351e682b347beae954c60ccf215edaf9c2da1abd4" Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.027358 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ngq69"] Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.034618 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-rlp2t"] Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.040672 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ngq69"] Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.046372 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-rlp2t"] Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.057264 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" path="/var/lib/kubelet/pods/0a6c0628-4522-4bb7-8a82-cc2e019eca2a/volumes" Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.058001 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" path="/var/lib/kubelet/pods/262733b9-4179-49e3-aee9-b62197cc89ba/volumes" Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.155930 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerStarted","Data":"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c"} Jan 21 11:37:02 crc kubenswrapper[4824]: I0121 11:37:02.170297 4824 generic.go:334] "Generic (PLEG): container finished" podID="5633f007-e1df-4760-b397-db8d77a2c43e" containerID="35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c" exitCode=0 Jan 21 11:37:02 crc kubenswrapper[4824]: I0121 11:37:02.170390 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerDied","Data":"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c"} Jan 21 11:37:03 crc kubenswrapper[4824]: I0121 11:37:03.179136 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerStarted","Data":"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31"} Jan 21 11:37:03 crc kubenswrapper[4824]: I0121 11:37:03.198284 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j5kv8" podStartSLOduration=2.641608016 podStartE2EDuration="7.198270088s" podCreationTimestamp="2026-01-21 11:36:56 +0000 UTC" firstStartedPulling="2026-01-21 11:36:58.145689972 +0000 UTC m=+1560.438719254" lastFinishedPulling="2026-01-21 11:37:02.702352034 +0000 UTC m=+1564.995381326" observedRunningTime="2026-01-21 11:37:03.198088274 +0000 UTC m=+1565.491117567" watchObservedRunningTime="2026-01-21 11:37:03.198270088 +0000 UTC m=+1565.491299380" Jan 21 11:37:06 crc kubenswrapper[4824]: I0121 11:37:06.808412 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:06 crc kubenswrapper[4824]: I0121 11:37:06.808640 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:06 crc kubenswrapper[4824]: I0121 11:37:06.841686 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:07 crc kubenswrapper[4824]: I0121 11:37:07.232541 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:07 crc kubenswrapper[4824]: I0121 11:37:07.271597 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.213939 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j5kv8" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="registry-server" containerID="cri-o://843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31" gracePeriod=2 Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.570372 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.637181 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rqjc\" (UniqueName: \"kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc\") pod \"5633f007-e1df-4760-b397-db8d77a2c43e\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.637277 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities\") pod \"5633f007-e1df-4760-b397-db8d77a2c43e\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.637448 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content\") pod \"5633f007-e1df-4760-b397-db8d77a2c43e\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.637912 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities" (OuterVolumeSpecName: "utilities") pod "5633f007-e1df-4760-b397-db8d77a2c43e" (UID: "5633f007-e1df-4760-b397-db8d77a2c43e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.641463 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc" (OuterVolumeSpecName: "kube-api-access-9rqjc") pod "5633f007-e1df-4760-b397-db8d77a2c43e" (UID: "5633f007-e1df-4760-b397-db8d77a2c43e"). InnerVolumeSpecName "kube-api-access-9rqjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.724568 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5633f007-e1df-4760-b397-db8d77a2c43e" (UID: "5633f007-e1df-4760-b397-db8d77a2c43e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.739006 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.739029 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rqjc\" (UniqueName: \"kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.739051 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.049776 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:37:10 crc kubenswrapper[4824]: E0121 11:37:10.050460 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.223198 4824 generic.go:334] "Generic (PLEG): container finished" podID="5633f007-e1df-4760-b397-db8d77a2c43e" containerID="843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31" exitCode=0 Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.223233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerDied","Data":"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31"} Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.223254 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerDied","Data":"9730f16fe571b01572a5f22435b85174e662bfa2a84a0568a751de97d620c25c"} Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.223268 4824 scope.go:117] "RemoveContainer" containerID="843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.223368 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.247566 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.254254 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.257692 4824 scope.go:117] "RemoveContainer" containerID="35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.274689 4824 scope.go:117] "RemoveContainer" containerID="e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.304873 4824 scope.go:117] "RemoveContainer" containerID="843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31" Jan 21 11:37:10 crc kubenswrapper[4824]: E0121 11:37:10.305366 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31\": container with ID starting with 843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31 not found: ID does not exist" containerID="843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.305399 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31"} err="failed to get container status \"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31\": rpc error: code = NotFound desc = could not find container \"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31\": container with ID starting with 843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31 not found: ID does not exist" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.305437 4824 scope.go:117] "RemoveContainer" containerID="35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c" Jan 21 11:37:10 crc kubenswrapper[4824]: E0121 11:37:10.305721 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c\": container with ID starting with 35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c not found: ID does not exist" containerID="35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.305746 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c"} err="failed to get container status \"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c\": rpc error: code = NotFound desc = could not find container \"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c\": container with ID starting with 35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c not found: ID does not exist" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.305760 4824 scope.go:117] "RemoveContainer" containerID="e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d" Jan 21 11:37:10 crc kubenswrapper[4824]: E0121 11:37:10.305939 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d\": container with ID starting with e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d not found: ID does not exist" containerID="e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.305976 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d"} err="failed to get container status \"e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d\": rpc error: code = NotFound desc = could not find container \"e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d\": container with ID starting with e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d not found: ID does not exist" Jan 21 11:37:12 crc kubenswrapper[4824]: I0121 11:37:12.057916 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" path="/var/lib/kubelet/pods/5633f007-e1df-4760-b397-db8d77a2c43e/volumes" Jan 21 11:37:15 crc kubenswrapper[4824]: I0121 11:37:15.258214 4824 generic.go:334] "Generic (PLEG): container finished" podID="60fedbf7-07e6-41be-ada3-2a06f28835e3" containerID="495175a3ec4152174c9a4e2a8aae14a46ac8d180433f99dc41b0fab262a48f7c" exitCode=0 Jan 21 11:37:15 crc kubenswrapper[4824]: I0121 11:37:15.258263 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" event={"ID":"60fedbf7-07e6-41be-ada3-2a06f28835e3","Type":"ContainerDied","Data":"495175a3ec4152174c9a4e2a8aae14a46ac8d180433f99dc41b0fab262a48f7c"} Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.561947 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.657053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory\") pod \"60fedbf7-07e6-41be-ada3-2a06f28835e3\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.657127 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam\") pod \"60fedbf7-07e6-41be-ada3-2a06f28835e3\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.657285 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h2jt\" (UniqueName: \"kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt\") pod \"60fedbf7-07e6-41be-ada3-2a06f28835e3\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.661437 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt" (OuterVolumeSpecName: "kube-api-access-7h2jt") pod "60fedbf7-07e6-41be-ada3-2a06f28835e3" (UID: "60fedbf7-07e6-41be-ada3-2a06f28835e3"). InnerVolumeSpecName "kube-api-access-7h2jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.677571 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory" (OuterVolumeSpecName: "inventory") pod "60fedbf7-07e6-41be-ada3-2a06f28835e3" (UID: "60fedbf7-07e6-41be-ada3-2a06f28835e3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.678212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "60fedbf7-07e6-41be-ada3-2a06f28835e3" (UID: "60fedbf7-07e6-41be-ada3-2a06f28835e3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.759058 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h2jt\" (UniqueName: \"kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.759096 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.759105 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.271373 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" event={"ID":"60fedbf7-07e6-41be-ada3-2a06f28835e3","Type":"ContainerDied","Data":"bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490"} Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.271645 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.271402 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.403634 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n"] Jan 21 11:37:17 crc kubenswrapper[4824]: E0121 11:37:17.404041 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="extract-utilities" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404058 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="extract-utilities" Jan 21 11:37:17 crc kubenswrapper[4824]: E0121 11:37:17.404081 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="extract-content" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404088 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="extract-content" Jan 21 11:37:17 crc kubenswrapper[4824]: E0121 11:37:17.404098 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60fedbf7-07e6-41be-ada3-2a06f28835e3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404106 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="60fedbf7-07e6-41be-ada3-2a06f28835e3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:17 crc kubenswrapper[4824]: E0121 11:37:17.404140 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="registry-server" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404146 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="registry-server" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404320 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="60fedbf7-07e6-41be-ada3-2a06f28835e3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404341 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="registry-server" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404873 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.406439 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.406745 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.406871 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.407095 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.409843 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n"] Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.469375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wqvn\" (UniqueName: \"kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.469442 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.469582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.571234 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.571399 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.571536 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wqvn\" (UniqueName: \"kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.575742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.576528 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.584665 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wqvn\" (UniqueName: \"kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.728516 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.176009 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n"] Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.280005 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" event={"ID":"8d46d1eb-4230-49c7-8376-698e7689ad2b","Type":"ContainerStarted","Data":"7c498c4871058e9b8aef05f8de415d1229252f8805909c3848b5ff58ff925394"} Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.551782 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.553577 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.561114 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.586948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.587079 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbn97\" (UniqueName: \"kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.587138 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.688405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.688507 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbn97\" (UniqueName: \"kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.688574 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.688924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.689169 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.704879 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbn97\" (UniqueName: \"kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.872690 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:19 crc kubenswrapper[4824]: I0121 11:37:19.288344 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" event={"ID":"8d46d1eb-4230-49c7-8376-698e7689ad2b","Type":"ContainerStarted","Data":"985d679df703288004096b440faf28c305917ee0a67a88fa7530222f8ee9d524"} Jan 21 11:37:19 crc kubenswrapper[4824]: I0121 11:37:19.303785 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" podStartSLOduration=1.8169619959999999 podStartE2EDuration="2.303770552s" podCreationTimestamp="2026-01-21 11:37:17 +0000 UTC" firstStartedPulling="2026-01-21 11:37:18.178888249 +0000 UTC m=+1580.471917541" lastFinishedPulling="2026-01-21 11:37:18.665696805 +0000 UTC m=+1580.958726097" observedRunningTime="2026-01-21 11:37:19.303537893 +0000 UTC m=+1581.596567185" watchObservedRunningTime="2026-01-21 11:37:19.303770552 +0000 UTC m=+1581.596799844" Jan 21 11:37:19 crc kubenswrapper[4824]: I0121 11:37:19.331024 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:19 crc kubenswrapper[4824]: W0121 11:37:19.332155 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16274e36_c051_4ef8_9120_07b9996f43d6.slice/crio-afc27a2346681f4afcd3cafb71b79b494dc2cff90b0fdb1ed7c606d5655d8650 WatchSource:0}: Error finding container afc27a2346681f4afcd3cafb71b79b494dc2cff90b0fdb1ed7c606d5655d8650: Status 404 returned error can't find the container with id afc27a2346681f4afcd3cafb71b79b494dc2cff90b0fdb1ed7c606d5655d8650 Jan 21 11:37:20 crc kubenswrapper[4824]: I0121 11:37:20.295488 4824 generic.go:334] "Generic (PLEG): container finished" podID="16274e36-c051-4ef8-9120-07b9996f43d6" containerID="e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e" exitCode=0 Jan 21 11:37:20 crc kubenswrapper[4824]: I0121 11:37:20.295925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerDied","Data":"e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e"} Jan 21 11:37:20 crc kubenswrapper[4824]: I0121 11:37:20.295987 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerStarted","Data":"afc27a2346681f4afcd3cafb71b79b494dc2cff90b0fdb1ed7c606d5655d8650"} Jan 21 11:37:21 crc kubenswrapper[4824]: I0121 11:37:21.303813 4824 generic.go:334] "Generic (PLEG): container finished" podID="16274e36-c051-4ef8-9120-07b9996f43d6" containerID="c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b" exitCode=0 Jan 21 11:37:21 crc kubenswrapper[4824]: I0121 11:37:21.303847 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerDied","Data":"c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b"} Jan 21 11:37:22 crc kubenswrapper[4824]: I0121 11:37:22.048585 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:37:22 crc kubenswrapper[4824]: E0121 11:37:22.049000 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:37:22 crc kubenswrapper[4824]: I0121 11:37:22.311728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerStarted","Data":"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff"} Jan 21 11:37:22 crc kubenswrapper[4824]: I0121 11:37:22.324545 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rpk5p" podStartSLOduration=2.827084818 podStartE2EDuration="4.324528962s" podCreationTimestamp="2026-01-21 11:37:18 +0000 UTC" firstStartedPulling="2026-01-21 11:37:20.297062348 +0000 UTC m=+1582.590091640" lastFinishedPulling="2026-01-21 11:37:21.794506493 +0000 UTC m=+1584.087535784" observedRunningTime="2026-01-21 11:37:22.322279104 +0000 UTC m=+1584.615308396" watchObservedRunningTime="2026-01-21 11:37:22.324528962 +0000 UTC m=+1584.617558254" Jan 21 11:37:28 crc kubenswrapper[4824]: I0121 11:37:28.873442 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:28 crc kubenswrapper[4824]: I0121 11:37:28.874083 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:28 crc kubenswrapper[4824]: I0121 11:37:28.911833 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:29 crc kubenswrapper[4824]: I0121 11:37:29.401837 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:29 crc kubenswrapper[4824]: I0121 11:37:29.442121 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:31 crc kubenswrapper[4824]: I0121 11:37:31.034978 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4h8mx"] Jan 21 11:37:31 crc kubenswrapper[4824]: I0121 11:37:31.044447 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-16cb-account-create-update-qjbx7"] Jan 21 11:37:31 crc kubenswrapper[4824]: I0121 11:37:31.052089 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-16cb-account-create-update-qjbx7"] Jan 21 11:37:31 crc kubenswrapper[4824]: I0121 11:37:31.060283 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4h8mx"] Jan 21 11:37:31 crc kubenswrapper[4824]: I0121 11:37:31.382212 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rpk5p" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="registry-server" containerID="cri-o://d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff" gracePeriod=2 Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.021989 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-91fe-account-create-update-4xgv2"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.029138 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-90fb-account-create-update-m4n4l"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.037057 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-dhvfz"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.042816 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-90fb-account-create-update-m4n4l"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.057255 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34afc8af-6ce5-4ee0-917b-326bd60d6237" path="/var/lib/kubelet/pods/34afc8af-6ce5-4ee0-917b-326bd60d6237/volumes" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.057985 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91261d07-56c8-4da3-8dd5-6f23587d77be" path="/var/lib/kubelet/pods/91261d07-56c8-4da3-8dd5-6f23587d77be/volumes" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.058703 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b24eefe0-9cb1-4672-85d8-6dfe2babebff" path="/var/lib/kubelet/pods/b24eefe0-9cb1-4672-85d8-6dfe2babebff/volumes" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.059312 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-91fe-account-create-update-4xgv2"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.059331 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-dhvfz"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.239705 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.350363 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities\") pod \"16274e36-c051-4ef8-9120-07b9996f43d6\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.350637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbn97\" (UniqueName: \"kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97\") pod \"16274e36-c051-4ef8-9120-07b9996f43d6\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.350783 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content\") pod \"16274e36-c051-4ef8-9120-07b9996f43d6\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.351181 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities" (OuterVolumeSpecName: "utilities") pod "16274e36-c051-4ef8-9120-07b9996f43d6" (UID: "16274e36-c051-4ef8-9120-07b9996f43d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.351412 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.355157 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97" (OuterVolumeSpecName: "kube-api-access-zbn97") pod "16274e36-c051-4ef8-9120-07b9996f43d6" (UID: "16274e36-c051-4ef8-9120-07b9996f43d6"). InnerVolumeSpecName "kube-api-access-zbn97". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.385777 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16274e36-c051-4ef8-9120-07b9996f43d6" (UID: "16274e36-c051-4ef8-9120-07b9996f43d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.392202 4824 generic.go:334] "Generic (PLEG): container finished" podID="16274e36-c051-4ef8-9120-07b9996f43d6" containerID="d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff" exitCode=0 Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.392258 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.392250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerDied","Data":"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff"} Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.392385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerDied","Data":"afc27a2346681f4afcd3cafb71b79b494dc2cff90b0fdb1ed7c606d5655d8650"} Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.392403 4824 scope.go:117] "RemoveContainer" containerID="d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.413832 4824 scope.go:117] "RemoveContainer" containerID="c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.420168 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.429227 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.444179 4824 scope.go:117] "RemoveContainer" containerID="e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.453540 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbn97\" (UniqueName: \"kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.453568 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.463719 4824 scope.go:117] "RemoveContainer" containerID="d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff" Jan 21 11:37:32 crc kubenswrapper[4824]: E0121 11:37:32.464083 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff\": container with ID starting with d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff not found: ID does not exist" containerID="d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.464128 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff"} err="failed to get container status \"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff\": rpc error: code = NotFound desc = could not find container \"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff\": container with ID starting with d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff not found: ID does not exist" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.464151 4824 scope.go:117] "RemoveContainer" containerID="c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b" Jan 21 11:37:32 crc kubenswrapper[4824]: E0121 11:37:32.464468 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b\": container with ID starting with c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b not found: ID does not exist" containerID="c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.464498 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b"} err="failed to get container status \"c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b\": rpc error: code = NotFound desc = could not find container \"c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b\": container with ID starting with c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b not found: ID does not exist" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.464529 4824 scope.go:117] "RemoveContainer" containerID="e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e" Jan 21 11:37:32 crc kubenswrapper[4824]: E0121 11:37:32.464767 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e\": container with ID starting with e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e not found: ID does not exist" containerID="e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.464797 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e"} err="failed to get container status \"e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e\": rpc error: code = NotFound desc = could not find container \"e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e\": container with ID starting with e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e not found: ID does not exist" Jan 21 11:37:33 crc kubenswrapper[4824]: I0121 11:37:33.021515 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-f2xfg"] Jan 21 11:37:33 crc kubenswrapper[4824]: I0121 11:37:33.029401 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-f2xfg"] Jan 21 11:37:33 crc kubenswrapper[4824]: I0121 11:37:33.049610 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:37:33 crc kubenswrapper[4824]: E0121 11:37:33.050443 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:37:34 crc kubenswrapper[4824]: I0121 11:37:34.056924 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12e7657f-78fe-4b5c-b614-4449e94366b9" path="/var/lib/kubelet/pods/12e7657f-78fe-4b5c-b614-4449e94366b9/volumes" Jan 21 11:37:34 crc kubenswrapper[4824]: I0121 11:37:34.057706 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" path="/var/lib/kubelet/pods/16274e36-c051-4ef8-9120-07b9996f43d6/volumes" Jan 21 11:37:34 crc kubenswrapper[4824]: I0121 11:37:34.058373 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60da7490-4aa1-4880-bfcb-e51a4ed99f91" path="/var/lib/kubelet/pods/60da7490-4aa1-4880-bfcb-e51a4ed99f91/volumes" Jan 21 11:37:34 crc kubenswrapper[4824]: I0121 11:37:34.059300 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5162c70-9359-42d5-a02d-16140ac4e366" path="/var/lib/kubelet/pods/d5162c70-9359-42d5-a02d-16140ac4e366/volumes" Jan 21 11:37:48 crc kubenswrapper[4824]: I0121 11:37:48.054136 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:37:48 crc kubenswrapper[4824]: E0121 11:37:48.054733 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:37:52 crc kubenswrapper[4824]: I0121 11:37:52.027162 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k6zm8"] Jan 21 11:37:52 crc kubenswrapper[4824]: I0121 11:37:52.033440 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k6zm8"] Jan 21 11:37:52 crc kubenswrapper[4824]: I0121 11:37:52.057211 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c73eda8-d376-4e18-9b33-f5ac4d10d091" path="/var/lib/kubelet/pods/9c73eda8-d376-4e18-9b33-f5ac4d10d091/volumes" Jan 21 11:37:52 crc kubenswrapper[4824]: I0121 11:37:52.525261 4824 generic.go:334] "Generic (PLEG): container finished" podID="8d46d1eb-4230-49c7-8376-698e7689ad2b" containerID="985d679df703288004096b440faf28c305917ee0a67a88fa7530222f8ee9d524" exitCode=0 Jan 21 11:37:52 crc kubenswrapper[4824]: I0121 11:37:52.525304 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" event={"ID":"8d46d1eb-4230-49c7-8376-698e7689ad2b","Type":"ContainerDied","Data":"985d679df703288004096b440faf28c305917ee0a67a88fa7530222f8ee9d524"} Jan 21 11:37:53 crc kubenswrapper[4824]: I0121 11:37:53.872410 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.032852 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wqvn\" (UniqueName: \"kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn\") pod \"8d46d1eb-4230-49c7-8376-698e7689ad2b\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.033083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam\") pod \"8d46d1eb-4230-49c7-8376-698e7689ad2b\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.033110 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory\") pod \"8d46d1eb-4230-49c7-8376-698e7689ad2b\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.037347 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn" (OuterVolumeSpecName: "kube-api-access-4wqvn") pod "8d46d1eb-4230-49c7-8376-698e7689ad2b" (UID: "8d46d1eb-4230-49c7-8376-698e7689ad2b"). InnerVolumeSpecName "kube-api-access-4wqvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.053813 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory" (OuterVolumeSpecName: "inventory") pod "8d46d1eb-4230-49c7-8376-698e7689ad2b" (UID: "8d46d1eb-4230-49c7-8376-698e7689ad2b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.054056 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8d46d1eb-4230-49c7-8376-698e7689ad2b" (UID: "8d46d1eb-4230-49c7-8376-698e7689ad2b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.135443 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.135467 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.135476 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wqvn\" (UniqueName: \"kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.538354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" event={"ID":"8d46d1eb-4230-49c7-8376-698e7689ad2b","Type":"ContainerDied","Data":"7c498c4871058e9b8aef05f8de415d1229252f8805909c3848b5ff58ff925394"} Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.538390 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c498c4871058e9b8aef05f8de415d1229252f8805909c3848b5ff58ff925394" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.538396 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.594857 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wbrwb"] Jan 21 11:37:54 crc kubenswrapper[4824]: E0121 11:37:54.595228 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="registry-server" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595247 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="registry-server" Jan 21 11:37:54 crc kubenswrapper[4824]: E0121 11:37:54.595269 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="extract-content" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595275 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="extract-content" Jan 21 11:37:54 crc kubenswrapper[4824]: E0121 11:37:54.595284 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="extract-utilities" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595290 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="extract-utilities" Jan 21 11:37:54 crc kubenswrapper[4824]: E0121 11:37:54.595298 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d46d1eb-4230-49c7-8376-698e7689ad2b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595305 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d46d1eb-4230-49c7-8376-698e7689ad2b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595461 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="registry-server" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595473 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d46d1eb-4230-49c7-8376-698e7689ad2b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.596023 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.597732 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.597784 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.597794 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.598290 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.608211 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wbrwb"] Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.744885 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.745105 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2wph\" (UniqueName: \"kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.745391 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.847289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2wph\" (UniqueName: \"kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.847435 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.847469 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.850197 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.850339 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.860271 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2wph\" (UniqueName: \"kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.909885 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:55 crc kubenswrapper[4824]: I0121 11:37:55.347142 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wbrwb"] Jan 21 11:37:55 crc kubenswrapper[4824]: I0121 11:37:55.545227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" event={"ID":"4529344f-3dfd-4858-83f1-4abffe47f2ad","Type":"ContainerStarted","Data":"e68763debb1695544c631b3afa9c054e02ec0885d12cf70e1432d634dc00418b"} Jan 21 11:37:56 crc kubenswrapper[4824]: I0121 11:37:56.552237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" event={"ID":"4529344f-3dfd-4858-83f1-4abffe47f2ad","Type":"ContainerStarted","Data":"39da814edb81d4f0345bc2ff3a9d25851f522015824a48410564608294c239ce"} Jan 21 11:37:56 crc kubenswrapper[4824]: I0121 11:37:56.566409 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" podStartSLOduration=2.02052861 podStartE2EDuration="2.566393808s" podCreationTimestamp="2026-01-21 11:37:54 +0000 UTC" firstStartedPulling="2026-01-21 11:37:55.35336279 +0000 UTC m=+1617.646392082" lastFinishedPulling="2026-01-21 11:37:55.899227988 +0000 UTC m=+1618.192257280" observedRunningTime="2026-01-21 11:37:56.563917693 +0000 UTC m=+1618.856946985" watchObservedRunningTime="2026-01-21 11:37:56.566393808 +0000 UTC m=+1618.859423100" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.157831 4824 scope.go:117] "RemoveContainer" containerID="6fa0076da1636efc5cc0178f55ca211d393f930b6c63c3d85cac4eedd9f2d288" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.182142 4824 scope.go:117] "RemoveContainer" containerID="a9bd46298904e9e7c4cccbdd1cecf51e20eaa4ef17b0e64f0ae86c7cca0ec79d" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.212348 4824 scope.go:117] "RemoveContainer" containerID="4c9c6f12e02388718e6b9fff3020bdecb9cf28f3ef2f988918e2aeb78dac7d68" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.249485 4824 scope.go:117] "RemoveContainer" containerID="b5f108624f476c15b99bcdcb7c95850ef2819568712069d3c2536bdcb6daa49b" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.285934 4824 scope.go:117] "RemoveContainer" containerID="c2bb2e93e558c6a9e9ce5d9ae43e9812275d36c293fd644336fd7e9fcc3ba6c0" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.315237 4824 scope.go:117] "RemoveContainer" containerID="7ab5f4aea14ccac0c6081bda9c7469c2c70bff8d5179b2221cca7a70044cada5" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.333150 4824 scope.go:117] "RemoveContainer" containerID="06b41897897bde2d466371c85738035a8ab9ee6530d1dfecdba6e71d57264bfe" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.349098 4824 scope.go:117] "RemoveContainer" containerID="7d5e936f6ab78e748abd4aabb374f95965e7eaaf526fd1a755b675eed79144ec" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.365314 4824 scope.go:117] "RemoveContainer" containerID="edd12500feb323e4dab204d7803bc5160bb50ef2b5116e5ba7686e65e4e155fc" Jan 21 11:38:01 crc kubenswrapper[4824]: I0121 11:38:01.589875 4824 generic.go:334] "Generic (PLEG): container finished" podID="4529344f-3dfd-4858-83f1-4abffe47f2ad" containerID="39da814edb81d4f0345bc2ff3a9d25851f522015824a48410564608294c239ce" exitCode=0 Jan 21 11:38:01 crc kubenswrapper[4824]: I0121 11:38:01.589948 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" event={"ID":"4529344f-3dfd-4858-83f1-4abffe47f2ad","Type":"ContainerDied","Data":"39da814edb81d4f0345bc2ff3a9d25851f522015824a48410564608294c239ce"} Jan 21 11:38:02 crc kubenswrapper[4824]: I0121 11:38:02.049031 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:38:02 crc kubenswrapper[4824]: E0121 11:38:02.049302 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:38:02 crc kubenswrapper[4824]: I0121 11:38:02.909214 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.079461 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam\") pod \"4529344f-3dfd-4858-83f1-4abffe47f2ad\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.079527 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0\") pod \"4529344f-3dfd-4858-83f1-4abffe47f2ad\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.079584 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2wph\" (UniqueName: \"kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph\") pod \"4529344f-3dfd-4858-83f1-4abffe47f2ad\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.083866 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph" (OuterVolumeSpecName: "kube-api-access-c2wph") pod "4529344f-3dfd-4858-83f1-4abffe47f2ad" (UID: "4529344f-3dfd-4858-83f1-4abffe47f2ad"). InnerVolumeSpecName "kube-api-access-c2wph". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.099833 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "4529344f-3dfd-4858-83f1-4abffe47f2ad" (UID: "4529344f-3dfd-4858-83f1-4abffe47f2ad"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.099927 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4529344f-3dfd-4858-83f1-4abffe47f2ad" (UID: "4529344f-3dfd-4858-83f1-4abffe47f2ad"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.181724 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.181751 4824 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.181761 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2wph\" (UniqueName: \"kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.604940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" event={"ID":"4529344f-3dfd-4858-83f1-4abffe47f2ad","Type":"ContainerDied","Data":"e68763debb1695544c631b3afa9c054e02ec0885d12cf70e1432d634dc00418b"} Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.604985 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.604992 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e68763debb1695544c631b3afa9c054e02ec0885d12cf70e1432d634dc00418b" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.659398 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw"] Jan 21 11:38:03 crc kubenswrapper[4824]: E0121 11:38:03.659787 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4529344f-3dfd-4858-83f1-4abffe47f2ad" containerName="ssh-known-hosts-edpm-deployment" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.659805 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4529344f-3dfd-4858-83f1-4abffe47f2ad" containerName="ssh-known-hosts-edpm-deployment" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.660003 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4529344f-3dfd-4858-83f1-4abffe47f2ad" containerName="ssh-known-hosts-edpm-deployment" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.660569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.662416 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.662668 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.662795 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.663634 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.675130 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw"] Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.792754 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.793129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.793159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mph2g\" (UniqueName: \"kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.894635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.894687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mph2g\" (UniqueName: \"kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.894775 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.897790 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.903546 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.908866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mph2g\" (UniqueName: \"kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.975840 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:04 crc kubenswrapper[4824]: I0121 11:38:04.391522 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw"] Jan 21 11:38:04 crc kubenswrapper[4824]: I0121 11:38:04.612248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" event={"ID":"ad533443-01b7-4860-9920-93ed67f6b52f","Type":"ContainerStarted","Data":"fb01de3abedf34fe5f459bd3c02fff34247e31b70d197937119d7945bb60ea39"} Jan 21 11:38:05 crc kubenswrapper[4824]: I0121 11:38:05.620903 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" event={"ID":"ad533443-01b7-4860-9920-93ed67f6b52f","Type":"ContainerStarted","Data":"946d859a672c7d1f509346749f7eeff11c24665737fbec52db44d7f634ff05ac"} Jan 21 11:38:05 crc kubenswrapper[4824]: I0121 11:38:05.640001 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" podStartSLOduration=2.098352483 podStartE2EDuration="2.639988266s" podCreationTimestamp="2026-01-21 11:38:03 +0000 UTC" firstStartedPulling="2026-01-21 11:38:04.398894178 +0000 UTC m=+1626.691923471" lastFinishedPulling="2026-01-21 11:38:04.940529962 +0000 UTC m=+1627.233559254" observedRunningTime="2026-01-21 11:38:05.635028412 +0000 UTC m=+1627.928057704" watchObservedRunningTime="2026-01-21 11:38:05.639988266 +0000 UTC m=+1627.933017557" Jan 21 11:38:10 crc kubenswrapper[4824]: I0121 11:38:10.663065 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad533443-01b7-4860-9920-93ed67f6b52f" containerID="946d859a672c7d1f509346749f7eeff11c24665737fbec52db44d7f634ff05ac" exitCode=0 Jan 21 11:38:10 crc kubenswrapper[4824]: I0121 11:38:10.663139 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" event={"ID":"ad533443-01b7-4860-9920-93ed67f6b52f","Type":"ContainerDied","Data":"946d859a672c7d1f509346749f7eeff11c24665737fbec52db44d7f634ff05ac"} Jan 21 11:38:11 crc kubenswrapper[4824]: I0121 11:38:11.981471 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.127035 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam\") pod \"ad533443-01b7-4860-9920-93ed67f6b52f\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.127232 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mph2g\" (UniqueName: \"kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g\") pod \"ad533443-01b7-4860-9920-93ed67f6b52f\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.127273 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory\") pod \"ad533443-01b7-4860-9920-93ed67f6b52f\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.131353 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g" (OuterVolumeSpecName: "kube-api-access-mph2g") pod "ad533443-01b7-4860-9920-93ed67f6b52f" (UID: "ad533443-01b7-4860-9920-93ed67f6b52f"). InnerVolumeSpecName "kube-api-access-mph2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.147453 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory" (OuterVolumeSpecName: "inventory") pod "ad533443-01b7-4860-9920-93ed67f6b52f" (UID: "ad533443-01b7-4860-9920-93ed67f6b52f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.152311 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ad533443-01b7-4860-9920-93ed67f6b52f" (UID: "ad533443-01b7-4860-9920-93ed67f6b52f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.228909 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mph2g\" (UniqueName: \"kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.228933 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.228943 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.677038 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" event={"ID":"ad533443-01b7-4860-9920-93ed67f6b52f","Type":"ContainerDied","Data":"fb01de3abedf34fe5f459bd3c02fff34247e31b70d197937119d7945bb60ea39"} Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.677065 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.677073 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb01de3abedf34fe5f459bd3c02fff34247e31b70d197937119d7945bb60ea39" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.722765 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w"] Jan 21 11:38:12 crc kubenswrapper[4824]: E0121 11:38:12.723125 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad533443-01b7-4860-9920-93ed67f6b52f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.723142 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad533443-01b7-4860-9920-93ed67f6b52f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.723370 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad533443-01b7-4860-9920-93ed67f6b52f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.723871 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.726731 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.726801 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.727084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.727116 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.730777 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w"] Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.735177 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.735469 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d25d6\" (UniqueName: \"kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.735633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.837246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.837330 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d25d6\" (UniqueName: \"kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.837552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.841431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.843083 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.851048 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d25d6\" (UniqueName: \"kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:13 crc kubenswrapper[4824]: I0121 11:38:13.036854 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:13 crc kubenswrapper[4824]: I0121 11:38:13.349595 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w"] Jan 21 11:38:13 crc kubenswrapper[4824]: I0121 11:38:13.684106 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" event={"ID":"890acdab-9f31-4afc-80e4-c1df308ac5a6","Type":"ContainerStarted","Data":"b95a36f4b2a1281221a694900f274ed364cc1352c3fe01992775f1762a0a247c"} Jan 21 11:38:14 crc kubenswrapper[4824]: I0121 11:38:14.021652 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9vq5"] Jan 21 11:38:14 crc kubenswrapper[4824]: I0121 11:38:14.027544 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9vq5"] Jan 21 11:38:14 crc kubenswrapper[4824]: I0121 11:38:14.057287 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baa34da9-3711-4cd4-a5eb-ee5814ac51a7" path="/var/lib/kubelet/pods/baa34da9-3711-4cd4-a5eb-ee5814ac51a7/volumes" Jan 21 11:38:14 crc kubenswrapper[4824]: I0121 11:38:14.691410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" event={"ID":"890acdab-9f31-4afc-80e4-c1df308ac5a6","Type":"ContainerStarted","Data":"882b40080daa5269b988b3f409290f0fb60c6467f759ffaa24e78d2d41b4111c"} Jan 21 11:38:14 crc kubenswrapper[4824]: I0121 11:38:14.706632 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" podStartSLOduration=2.190013548 podStartE2EDuration="2.706616202s" podCreationTimestamp="2026-01-21 11:38:12 +0000 UTC" firstStartedPulling="2026-01-21 11:38:13.350115932 +0000 UTC m=+1635.643145224" lastFinishedPulling="2026-01-21 11:38:13.866718587 +0000 UTC m=+1636.159747878" observedRunningTime="2026-01-21 11:38:14.706190348 +0000 UTC m=+1636.999219640" watchObservedRunningTime="2026-01-21 11:38:14.706616202 +0000 UTC m=+1636.999645494" Jan 21 11:38:15 crc kubenswrapper[4824]: I0121 11:38:15.020304 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7gtcg"] Jan 21 11:38:15 crc kubenswrapper[4824]: I0121 11:38:15.026822 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7gtcg"] Jan 21 11:38:16 crc kubenswrapper[4824]: I0121 11:38:16.057006 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="135a8573-b55a-4c5f-9cb2-a7c3adea9720" path="/var/lib/kubelet/pods/135a8573-b55a-4c5f-9cb2-a7c3adea9720/volumes" Jan 21 11:38:17 crc kubenswrapper[4824]: I0121 11:38:17.049445 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:38:17 crc kubenswrapper[4824]: E0121 11:38:17.049685 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:38:20 crc kubenswrapper[4824]: I0121 11:38:20.729176 4824 generic.go:334] "Generic (PLEG): container finished" podID="890acdab-9f31-4afc-80e4-c1df308ac5a6" containerID="882b40080daa5269b988b3f409290f0fb60c6467f759ffaa24e78d2d41b4111c" exitCode=0 Jan 21 11:38:20 crc kubenswrapper[4824]: I0121 11:38:20.729254 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" event={"ID":"890acdab-9f31-4afc-80e4-c1df308ac5a6","Type":"ContainerDied","Data":"882b40080daa5269b988b3f409290f0fb60c6467f759ffaa24e78d2d41b4111c"} Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.056575 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.098169 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d25d6\" (UniqueName: \"kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6\") pod \"890acdab-9f31-4afc-80e4-c1df308ac5a6\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.098213 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory\") pod \"890acdab-9f31-4afc-80e4-c1df308ac5a6\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.098247 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam\") pod \"890acdab-9f31-4afc-80e4-c1df308ac5a6\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.102509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6" (OuterVolumeSpecName: "kube-api-access-d25d6") pod "890acdab-9f31-4afc-80e4-c1df308ac5a6" (UID: "890acdab-9f31-4afc-80e4-c1df308ac5a6"). InnerVolumeSpecName "kube-api-access-d25d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.119061 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "890acdab-9f31-4afc-80e4-c1df308ac5a6" (UID: "890acdab-9f31-4afc-80e4-c1df308ac5a6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.119107 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory" (OuterVolumeSpecName: "inventory") pod "890acdab-9f31-4afc-80e4-c1df308ac5a6" (UID: "890acdab-9f31-4afc-80e4-c1df308ac5a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.200609 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d25d6\" (UniqueName: \"kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.200633 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.200643 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.741283 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" event={"ID":"890acdab-9f31-4afc-80e4-c1df308ac5a6","Type":"ContainerDied","Data":"b95a36f4b2a1281221a694900f274ed364cc1352c3fe01992775f1762a0a247c"} Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.741316 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b95a36f4b2a1281221a694900f274ed364cc1352c3fe01992775f1762a0a247c" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.741533 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.792499 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2"] Jan 21 11:38:22 crc kubenswrapper[4824]: E0121 11:38:22.792847 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="890acdab-9f31-4afc-80e4-c1df308ac5a6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.792866 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="890acdab-9f31-4afc-80e4-c1df308ac5a6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.793046 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="890acdab-9f31-4afc-80e4-c1df308ac5a6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.793580 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.796523 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.796901 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.796913 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.797239 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.797289 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.797334 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.797350 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.797586 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.804059 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2"] Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807091 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807339 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807461 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807785 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhxkm\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.808089 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.808339 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.808380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.808412 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.808457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.909726 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.909910 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910375 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910460 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910535 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910613 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910791 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910864 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhxkm\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.911035 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.913571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.913586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.913900 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.914040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.914196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.914300 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.914591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.914600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.915599 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.915847 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.916582 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.917428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.917538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.926031 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhxkm\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:23 crc kubenswrapper[4824]: I0121 11:38:23.109050 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:23 crc kubenswrapper[4824]: I0121 11:38:23.513654 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2"] Jan 21 11:38:23 crc kubenswrapper[4824]: I0121 11:38:23.748477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" event={"ID":"b821ea8d-9327-4623-b5be-3da7d3872cab","Type":"ContainerStarted","Data":"b432a7ecf664fae2f795f90960df62d71c3a7f5305cc1480692283386292bf1c"} Jan 21 11:38:24 crc kubenswrapper[4824]: I0121 11:38:24.757327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" event={"ID":"b821ea8d-9327-4623-b5be-3da7d3872cab","Type":"ContainerStarted","Data":"4463df9b5218899c21b0867a3bc9c0284f5c2175ab91fef1df2e89365b5d4cc1"} Jan 21 11:38:24 crc kubenswrapper[4824]: I0121 11:38:24.771632 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" podStartSLOduration=2.2999615970000002 podStartE2EDuration="2.7716189s" podCreationTimestamp="2026-01-21 11:38:22 +0000 UTC" firstStartedPulling="2026-01-21 11:38:23.515542889 +0000 UTC m=+1645.808572181" lastFinishedPulling="2026-01-21 11:38:23.987200192 +0000 UTC m=+1646.280229484" observedRunningTime="2026-01-21 11:38:24.769013481 +0000 UTC m=+1647.062042774" watchObservedRunningTime="2026-01-21 11:38:24.7716189 +0000 UTC m=+1647.064648192" Jan 21 11:38:29 crc kubenswrapper[4824]: I0121 11:38:29.050006 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:38:29 crc kubenswrapper[4824]: E0121 11:38:29.051274 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:38:43 crc kubenswrapper[4824]: I0121 11:38:43.048982 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:38:43 crc kubenswrapper[4824]: E0121 11:38:43.049735 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:38:50 crc kubenswrapper[4824]: I0121 11:38:50.935288 4824 generic.go:334] "Generic (PLEG): container finished" podID="b821ea8d-9327-4623-b5be-3da7d3872cab" containerID="4463df9b5218899c21b0867a3bc9c0284f5c2175ab91fef1df2e89365b5d4cc1" exitCode=0 Jan 21 11:38:50 crc kubenswrapper[4824]: I0121 11:38:50.935369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" event={"ID":"b821ea8d-9327-4623-b5be-3da7d3872cab","Type":"ContainerDied","Data":"4463df9b5218899c21b0867a3bc9c0284f5c2175ab91fef1df2e89365b5d4cc1"} Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.245125 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335216 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335370 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335385 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335401 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335417 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335468 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335525 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335569 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335605 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335644 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhxkm\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335684 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335704 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.340455 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.340778 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341229 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341354 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341496 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm" (OuterVolumeSpecName: "kube-api-access-dhxkm") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "kube-api-access-dhxkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341720 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341739 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341755 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.342640 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.343289 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.343656 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: E0121 11:38:52.355863 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam podName:b821ea8d-9327-4623-b5be-3da7d3872cab nodeName:}" failed. No retries permitted until 2026-01-21 11:38:52.855840133 +0000 UTC m=+1675.148869425 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key-openstack-edpm-ipam" (UniqueName: "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab") : error deleting /var/lib/kubelet/pods/b821ea8d-9327-4623-b5be-3da7d3872cab/volume-subpaths: remove /var/lib/kubelet/pods/b821ea8d-9327-4623-b5be-3da7d3872cab/volume-subpaths: no such file or directory Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.358470 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory" (OuterVolumeSpecName: "inventory") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438088 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438116 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438129 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438140 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438151 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438161 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438169 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438177 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438185 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438193 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhxkm\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438201 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438211 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438220 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.945864 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.950128 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.954384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" event={"ID":"b821ea8d-9327-4623-b5be-3da7d3872cab","Type":"ContainerDied","Data":"b432a7ecf664fae2f795f90960df62d71c3a7f5305cc1480692283386292bf1c"} Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.954427 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b432a7ecf664fae2f795f90960df62d71c3a7f5305cc1480692283386292bf1c" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.954481 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.031319 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq"] Jan 21 11:38:53 crc kubenswrapper[4824]: E0121 11:38:53.031845 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b821ea8d-9327-4623-b5be-3da7d3872cab" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.031863 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b821ea8d-9327-4623-b5be-3da7d3872cab" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.032075 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b821ea8d-9327-4623-b5be-3da7d3872cab" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.032635 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.034205 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.034434 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.035416 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.035687 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.037082 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.038700 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq"] Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.052485 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.153664 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.154002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.154071 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7w4\" (UniqueName: \"kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.154147 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.154200 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.256112 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.256181 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.256228 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.256455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.256503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md7w4\" (UniqueName: \"kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.257329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.259707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.259918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.260917 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.270990 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7w4\" (UniqueName: \"kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.347829 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.770789 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq"] Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.967011 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" event={"ID":"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f","Type":"ContainerStarted","Data":"6d455d5215fcbfd8a47efb176dd7eb21ea2bd57fe6e56abd130fecb774e16899"} Jan 21 11:38:54 crc kubenswrapper[4824]: I0121 11:38:54.973602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" event={"ID":"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f","Type":"ContainerStarted","Data":"2ed717f75906d3a0f34ff5910051eb7ff1d604b1c193192f6557d69d614d1475"} Jan 21 11:38:54 crc kubenswrapper[4824]: I0121 11:38:54.990558 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" podStartSLOduration=1.47454299 podStartE2EDuration="1.990539079s" podCreationTimestamp="2026-01-21 11:38:53 +0000 UTC" firstStartedPulling="2026-01-21 11:38:53.773767981 +0000 UTC m=+1676.066797273" lastFinishedPulling="2026-01-21 11:38:54.28976407 +0000 UTC m=+1676.582793362" observedRunningTime="2026-01-21 11:38:54.98405149 +0000 UTC m=+1677.277080783" watchObservedRunningTime="2026-01-21 11:38:54.990539079 +0000 UTC m=+1677.283568372" Jan 21 11:38:56 crc kubenswrapper[4824]: I0121 11:38:56.049375 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:38:56 crc kubenswrapper[4824]: E0121 11:38:56.050142 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:38:58 crc kubenswrapper[4824]: I0121 11:38:58.031508 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-6c642"] Jan 21 11:38:58 crc kubenswrapper[4824]: I0121 11:38:58.038230 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-6c642"] Jan 21 11:38:58 crc kubenswrapper[4824]: I0121 11:38:58.057162 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c162ff-82c2-4166-8a44-92e0455cae39" path="/var/lib/kubelet/pods/12c162ff-82c2-4166-8a44-92e0455cae39/volumes" Jan 21 11:39:00 crc kubenswrapper[4824]: I0121 11:39:00.512207 4824 scope.go:117] "RemoveContainer" containerID="e6d7bd7f99afa7be3d7eee7ba3147eedee78587b76d733c8cb546689d1b785fe" Jan 21 11:39:00 crc kubenswrapper[4824]: I0121 11:39:00.553468 4824 scope.go:117] "RemoveContainer" containerID="dd76c88e84c244299a1fbb85cad5296bd9ab946000d0d268303768f60e22c9c4" Jan 21 11:39:00 crc kubenswrapper[4824]: I0121 11:39:00.587831 4824 scope.go:117] "RemoveContainer" containerID="7468f8bf7c17d51a36aabd9a1f3d9c19a0cb1f2b474c4bfcd676598418dd86f7" Jan 21 11:39:07 crc kubenswrapper[4824]: I0121 11:39:07.048913 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:39:07 crc kubenswrapper[4824]: E0121 11:39:07.049754 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:39:18 crc kubenswrapper[4824]: I0121 11:39:18.053531 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:39:18 crc kubenswrapper[4824]: E0121 11:39:18.054245 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:39:29 crc kubenswrapper[4824]: I0121 11:39:29.049831 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:39:29 crc kubenswrapper[4824]: E0121 11:39:29.050990 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:39:38 crc kubenswrapper[4824]: I0121 11:39:38.244584 4824 generic.go:334] "Generic (PLEG): container finished" podID="18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" containerID="2ed717f75906d3a0f34ff5910051eb7ff1d604b1c193192f6557d69d614d1475" exitCode=0 Jan 21 11:39:38 crc kubenswrapper[4824]: I0121 11:39:38.244678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" event={"ID":"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f","Type":"ContainerDied","Data":"2ed717f75906d3a0f34ff5910051eb7ff1d604b1c193192f6557d69d614d1475"} Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.566766 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.661907 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory\") pod \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.661984 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle\") pod \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.662142 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0\") pod \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.662181 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md7w4\" (UniqueName: \"kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4\") pod \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.662205 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam\") pod \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.666564 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4" (OuterVolumeSpecName: "kube-api-access-md7w4") pod "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" (UID: "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f"). InnerVolumeSpecName "kube-api-access-md7w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.666600 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" (UID: "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.681257 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" (UID: "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.682373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory" (OuterVolumeSpecName: "inventory") pod "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" (UID: "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.682932 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" (UID: "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.764933 4824 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.764977 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md7w4\" (UniqueName: \"kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4\") on node \"crc\" DevicePath \"\"" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.764988 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.764998 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.765006 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.259762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" event={"ID":"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f","Type":"ContainerDied","Data":"6d455d5215fcbfd8a47efb176dd7eb21ea2bd57fe6e56abd130fecb774e16899"} Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.259801 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d455d5215fcbfd8a47efb176dd7eb21ea2bd57fe6e56abd130fecb774e16899" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.259845 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.323204 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l"] Jan 21 11:39:40 crc kubenswrapper[4824]: E0121 11:39:40.323595 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.323614 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.323785 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.324353 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.326357 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.326647 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.326691 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.326656 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.327301 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.329664 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.331307 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l"] Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375231 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375322 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw25j\" (UniqueName: \"kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375631 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.476765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw25j\" (UniqueName: \"kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.476835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.476932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.477599 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.477707 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.477819 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.481030 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.481263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.482355 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.484082 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.484156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.491536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw25j\" (UniqueName: \"kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.637313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:41 crc kubenswrapper[4824]: I0121 11:39:41.048912 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:39:41 crc kubenswrapper[4824]: E0121 11:39:41.049527 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:39:41 crc kubenswrapper[4824]: I0121 11:39:41.082811 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l"] Jan 21 11:39:41 crc kubenswrapper[4824]: I0121 11:39:41.267198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" event={"ID":"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d","Type":"ContainerStarted","Data":"ca1530bd513a46b192d4e6e7ec5ad68d0b40713a10661b674e5682caf2459e0c"} Jan 21 11:39:42 crc kubenswrapper[4824]: I0121 11:39:42.275126 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" event={"ID":"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d","Type":"ContainerStarted","Data":"6d5206d4d635a0ba105e072dcc9351ce51a4045b38ec07cb4ebf2776319165b3"} Jan 21 11:39:42 crc kubenswrapper[4824]: I0121 11:39:42.289212 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" podStartSLOduration=1.729008087 podStartE2EDuration="2.289198968s" podCreationTimestamp="2026-01-21 11:39:40 +0000 UTC" firstStartedPulling="2026-01-21 11:39:41.086063237 +0000 UTC m=+1723.379092529" lastFinishedPulling="2026-01-21 11:39:41.646254118 +0000 UTC m=+1723.939283410" observedRunningTime="2026-01-21 11:39:42.285723546 +0000 UTC m=+1724.578752837" watchObservedRunningTime="2026-01-21 11:39:42.289198968 +0000 UTC m=+1724.582228260" Jan 21 11:39:55 crc kubenswrapper[4824]: I0121 11:39:55.049939 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:39:55 crc kubenswrapper[4824]: E0121 11:39:55.050658 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:40:09 crc kubenswrapper[4824]: I0121 11:40:09.049137 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:40:09 crc kubenswrapper[4824]: E0121 11:40:09.050084 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:40:16 crc kubenswrapper[4824]: I0121 11:40:16.481908 4824 generic.go:334] "Generic (PLEG): container finished" podID="4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" containerID="6d5206d4d635a0ba105e072dcc9351ce51a4045b38ec07cb4ebf2776319165b3" exitCode=0 Jan 21 11:40:16 crc kubenswrapper[4824]: I0121 11:40:16.481988 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" event={"ID":"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d","Type":"ContainerDied","Data":"6d5206d4d635a0ba105e072dcc9351ce51a4045b38ec07cb4ebf2776319165b3"} Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.789128 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910150 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910212 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910318 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910396 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw25j\" (UniqueName: \"kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910442 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.914619 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j" (OuterVolumeSpecName: "kube-api-access-zw25j") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "kube-api-access-zw25j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.918916 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.930810 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.931034 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.932111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.932685 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory" (OuterVolumeSpecName: "inventory") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013136 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013165 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013175 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013185 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw25j\" (UniqueName: \"kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013194 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013205 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.496409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" event={"ID":"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d","Type":"ContainerDied","Data":"ca1530bd513a46b192d4e6e7ec5ad68d0b40713a10661b674e5682caf2459e0c"} Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.496444 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca1530bd513a46b192d4e6e7ec5ad68d0b40713a10661b674e5682caf2459e0c" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.496663 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.562483 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn"] Jan 21 11:40:18 crc kubenswrapper[4824]: E0121 11:40:18.562819 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.562836 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.563035 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.563529 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.565289 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.565307 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.566526 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.567144 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.567689 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.577432 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn"] Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.725795 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnnln\" (UniqueName: \"kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.725848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.726025 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.726371 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.726540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.828434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.828551 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.828691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnnln\" (UniqueName: \"kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.828721 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.828789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.831580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.831907 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.832485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.833129 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.841643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnnln\" (UniqueName: \"kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.877485 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:19 crc kubenswrapper[4824]: I0121 11:40:19.305813 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn"] Jan 21 11:40:19 crc kubenswrapper[4824]: W0121 11:40:19.306374 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdde8991_c204_45a0_b344_dcdc41a9a275.slice/crio-bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202 WatchSource:0}: Error finding container bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202: Status 404 returned error can't find the container with id bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202 Jan 21 11:40:19 crc kubenswrapper[4824]: I0121 11:40:19.503811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" event={"ID":"fdde8991-c204-45a0-b344-dcdc41a9a275","Type":"ContainerStarted","Data":"bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202"} Jan 21 11:40:20 crc kubenswrapper[4824]: I0121 11:40:20.511203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" event={"ID":"fdde8991-c204-45a0-b344-dcdc41a9a275","Type":"ContainerStarted","Data":"1f9e141178592276bca51a00e34332a99986e41de0caf5feaec25186ff7878f9"} Jan 21 11:40:20 crc kubenswrapper[4824]: I0121 11:40:20.522800 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" podStartSLOduration=2.044140647 podStartE2EDuration="2.522788581s" podCreationTimestamp="2026-01-21 11:40:18 +0000 UTC" firstStartedPulling="2026-01-21 11:40:19.308019365 +0000 UTC m=+1761.601048657" lastFinishedPulling="2026-01-21 11:40:19.7866673 +0000 UTC m=+1762.079696591" observedRunningTime="2026-01-21 11:40:20.52187017 +0000 UTC m=+1762.814899462" watchObservedRunningTime="2026-01-21 11:40:20.522788581 +0000 UTC m=+1762.815817874" Jan 21 11:40:21 crc kubenswrapper[4824]: I0121 11:40:21.049780 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:40:21 crc kubenswrapper[4824]: E0121 11:40:21.050223 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:40:34 crc kubenswrapper[4824]: I0121 11:40:34.050587 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:40:34 crc kubenswrapper[4824]: E0121 11:40:34.051348 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:40:47 crc kubenswrapper[4824]: I0121 11:40:47.048827 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:40:47 crc kubenswrapper[4824]: E0121 11:40:47.049393 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:41:00 crc kubenswrapper[4824]: I0121 11:41:00.049806 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:41:00 crc kubenswrapper[4824]: E0121 11:41:00.050423 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:41:13 crc kubenswrapper[4824]: I0121 11:41:13.049334 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:41:13 crc kubenswrapper[4824]: E0121 11:41:13.050110 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:41:26 crc kubenswrapper[4824]: I0121 11:41:26.049872 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:41:26 crc kubenswrapper[4824]: I0121 11:41:26.928833 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b"} Jan 21 11:43:10 crc kubenswrapper[4824]: I0121 11:43:10.589800 4824 generic.go:334] "Generic (PLEG): container finished" podID="fdde8991-c204-45a0-b344-dcdc41a9a275" containerID="1f9e141178592276bca51a00e34332a99986e41de0caf5feaec25186ff7878f9" exitCode=0 Jan 21 11:43:10 crc kubenswrapper[4824]: I0121 11:43:10.590529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" event={"ID":"fdde8991-c204-45a0-b344-dcdc41a9a275","Type":"ContainerDied","Data":"1f9e141178592276bca51a00e34332a99986e41de0caf5feaec25186ff7878f9"} Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.930472 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.969389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory\") pod \"fdde8991-c204-45a0-b344-dcdc41a9a275\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.969568 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam\") pod \"fdde8991-c204-45a0-b344-dcdc41a9a275\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.969654 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnnln\" (UniqueName: \"kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln\") pod \"fdde8991-c204-45a0-b344-dcdc41a9a275\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.969737 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0\") pod \"fdde8991-c204-45a0-b344-dcdc41a9a275\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.969802 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle\") pod \"fdde8991-c204-45a0-b344-dcdc41a9a275\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.976119 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "fdde8991-c204-45a0-b344-dcdc41a9a275" (UID: "fdde8991-c204-45a0-b344-dcdc41a9a275"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.979510 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln" (OuterVolumeSpecName: "kube-api-access-qnnln") pod "fdde8991-c204-45a0-b344-dcdc41a9a275" (UID: "fdde8991-c204-45a0-b344-dcdc41a9a275"). InnerVolumeSpecName "kube-api-access-qnnln". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.994168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory" (OuterVolumeSpecName: "inventory") pod "fdde8991-c204-45a0-b344-dcdc41a9a275" (UID: "fdde8991-c204-45a0-b344-dcdc41a9a275"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.994217 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "fdde8991-c204-45a0-b344-dcdc41a9a275" (UID: "fdde8991-c204-45a0-b344-dcdc41a9a275"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.000377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "fdde8991-c204-45a0-b344-dcdc41a9a275" (UID: "fdde8991-c204-45a0-b344-dcdc41a9a275"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.072977 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnnln\" (UniqueName: \"kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln\") on node \"crc\" DevicePath \"\"" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.073010 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.073019 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.073029 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.073040 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.605701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" event={"ID":"fdde8991-c204-45a0-b344-dcdc41a9a275","Type":"ContainerDied","Data":"bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202"} Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.605751 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.605757 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.674199 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8"] Jan 21 11:43:12 crc kubenswrapper[4824]: E0121 11:43:12.674596 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdde8991-c204-45a0-b344-dcdc41a9a275" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.674619 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdde8991-c204-45a0-b344-dcdc41a9a275" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.674816 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdde8991-c204-45a0-b344-dcdc41a9a275" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.675406 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.678465 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.678767 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.678982 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.679047 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.679174 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.679311 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.679435 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.683224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.683521 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.683687 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.683793 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.683862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.684172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.684316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.684530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.684570 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2hm6\" (UniqueName: \"kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.687819 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8"] Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.787797 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788265 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2hm6\" (UniqueName: \"kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788484 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788797 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.789530 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.791898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.792184 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.792444 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.793450 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.794545 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.794946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.795399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.801415 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2hm6\" (UniqueName: \"kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.993250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:13 crc kubenswrapper[4824]: I0121 11:43:13.436716 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8"] Jan 21 11:43:13 crc kubenswrapper[4824]: I0121 11:43:13.450251 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:43:13 crc kubenswrapper[4824]: I0121 11:43:13.614756 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" event={"ID":"7355a8fc-8c18-4de1-8baf-f6294ea9e11c","Type":"ContainerStarted","Data":"800a2f48f970dd7813b9b6aa1e55ec8212739dafcfa774a9c1c44f5bcec499fb"} Jan 21 11:43:14 crc kubenswrapper[4824]: I0121 11:43:14.621845 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" event={"ID":"7355a8fc-8c18-4de1-8baf-f6294ea9e11c","Type":"ContainerStarted","Data":"ffa277a7594795b5b30433a2fa1d40a85f0bccc5efe3c5c45b34f69bd8dc1a6d"} Jan 21 11:43:14 crc kubenswrapper[4824]: I0121 11:43:14.639371 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" podStartSLOduration=2.191622291 podStartE2EDuration="2.639357851s" podCreationTimestamp="2026-01-21 11:43:12 +0000 UTC" firstStartedPulling="2026-01-21 11:43:13.450026727 +0000 UTC m=+1935.743056020" lastFinishedPulling="2026-01-21 11:43:13.897762298 +0000 UTC m=+1936.190791580" observedRunningTime="2026-01-21 11:43:14.632796099 +0000 UTC m=+1936.925825391" watchObservedRunningTime="2026-01-21 11:43:14.639357851 +0000 UTC m=+1936.932387143" Jan 21 11:43:46 crc kubenswrapper[4824]: I0121 11:43:46.065216 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:43:46 crc kubenswrapper[4824]: I0121 11:43:46.065609 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:44:16 crc kubenswrapper[4824]: I0121 11:44:16.065326 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:44:16 crc kubenswrapper[4824]: I0121 11:44:16.065735 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.065753 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.066334 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.066370 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.067007 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.067054 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b" gracePeriod=600 Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.209887 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b" exitCode=0 Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.209974 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b"} Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.210166 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:44:47 crc kubenswrapper[4824]: I0121 11:44:47.217757 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788"} Jan 21 11:44:57 crc kubenswrapper[4824]: I0121 11:44:57.287685 4824 generic.go:334] "Generic (PLEG): container finished" podID="7355a8fc-8c18-4de1-8baf-f6294ea9e11c" containerID="ffa277a7594795b5b30433a2fa1d40a85f0bccc5efe3c5c45b34f69bd8dc1a6d" exitCode=0 Jan 21 11:44:57 crc kubenswrapper[4824]: I0121 11:44:57.287773 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" event={"ID":"7355a8fc-8c18-4de1-8baf-f6294ea9e11c","Type":"ContainerDied","Data":"ffa277a7594795b5b30433a2fa1d40a85f0bccc5efe3c5c45b34f69bd8dc1a6d"} Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.597288 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755048 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755101 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755165 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755182 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755265 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755295 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755372 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2hm6\" (UniqueName: \"kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755408 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.760738 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6" (OuterVolumeSpecName: "kube-api-access-t2hm6") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "kube-api-access-t2hm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.760341 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.776080 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.777408 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.778018 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory" (OuterVolumeSpecName: "inventory") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.778426 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.779007 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.779029 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.781609 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857433 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857650 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857661 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2hm6\" (UniqueName: \"kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857671 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857679 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857686 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857694 4824 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857701 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857709 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.300785 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" event={"ID":"7355a8fc-8c18-4de1-8baf-f6294ea9e11c","Type":"ContainerDied","Data":"800a2f48f970dd7813b9b6aa1e55ec8212739dafcfa774a9c1c44f5bcec499fb"} Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.300815 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.300822 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="800a2f48f970dd7813b9b6aa1e55ec8212739dafcfa774a9c1c44f5bcec499fb" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.412339 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6"] Jan 21 11:44:59 crc kubenswrapper[4824]: E0121 11:44:59.412852 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7355a8fc-8c18-4de1-8baf-f6294ea9e11c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.412938 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7355a8fc-8c18-4de1-8baf-f6294ea9e11c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.413219 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7355a8fc-8c18-4de1-8baf-f6294ea9e11c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.413771 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.417016 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.417136 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.417525 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.417777 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.417939 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.425029 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6"] Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569258 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569370 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgqg9\" (UniqueName: \"kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569488 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671281 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671396 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgqg9\" (UniqueName: \"kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671444 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671644 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.674626 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.674718 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.677455 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.677551 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.677578 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.677838 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.684773 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgqg9\" (UniqueName: \"kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.729565 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.129684 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh"] Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.130974 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.132755 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.133188 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.139645 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh"] Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.233818 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6"] Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.282759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.283050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.283329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgsr7\" (UniqueName: \"kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.307976 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" event={"ID":"49b2a136-e5c3-4741-80c6-4c545d4a82a0","Type":"ContainerStarted","Data":"b8c227627e16ce045a124bfbe36c3ecb09476a5cdfcf39963427bafa05e52cee"} Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.384894 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgsr7\" (UniqueName: \"kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.384986 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.385082 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.385913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.389803 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.398041 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgsr7\" (UniqueName: \"kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.450311 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.817730 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh"] Jan 21 11:45:00 crc kubenswrapper[4824]: W0121 11:45:00.819586 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35e40551_5ca0_42ef_8305_66fd69d36a9f.slice/crio-77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310 WatchSource:0}: Error finding container 77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310: Status 404 returned error can't find the container with id 77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310 Jan 21 11:45:01 crc kubenswrapper[4824]: I0121 11:45:01.316128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" event={"ID":"49b2a136-e5c3-4741-80c6-4c545d4a82a0","Type":"ContainerStarted","Data":"6cc48beb53c571878821f1833c1a75164e55e0ffea7eb2292d2abd90321cb03f"} Jan 21 11:45:01 crc kubenswrapper[4824]: I0121 11:45:01.317649 4824 generic.go:334] "Generic (PLEG): container finished" podID="35e40551-5ca0-42ef-8305-66fd69d36a9f" containerID="34d65e26d2a1896dfa268eb4b192173adbc67325a4a1d10c57a0833b00d02595" exitCode=0 Jan 21 11:45:01 crc kubenswrapper[4824]: I0121 11:45:01.317682 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" event={"ID":"35e40551-5ca0-42ef-8305-66fd69d36a9f","Type":"ContainerDied","Data":"34d65e26d2a1896dfa268eb4b192173adbc67325a4a1d10c57a0833b00d02595"} Jan 21 11:45:01 crc kubenswrapper[4824]: I0121 11:45:01.317711 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" event={"ID":"35e40551-5ca0-42ef-8305-66fd69d36a9f","Type":"ContainerStarted","Data":"77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310"} Jan 21 11:45:01 crc kubenswrapper[4824]: I0121 11:45:01.330536 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" podStartSLOduration=1.820143318 podStartE2EDuration="2.330522137s" podCreationTimestamp="2026-01-21 11:44:59 +0000 UTC" firstStartedPulling="2026-01-21 11:45:00.24527848 +0000 UTC m=+2042.538307773" lastFinishedPulling="2026-01-21 11:45:00.75565731 +0000 UTC m=+2043.048686592" observedRunningTime="2026-01-21 11:45:01.327297198 +0000 UTC m=+2043.620326490" watchObservedRunningTime="2026-01-21 11:45:01.330522137 +0000 UTC m=+2043.623551429" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.574264 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.720652 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume\") pod \"35e40551-5ca0-42ef-8305-66fd69d36a9f\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.720781 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume\") pod \"35e40551-5ca0-42ef-8305-66fd69d36a9f\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.720817 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgsr7\" (UniqueName: \"kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7\") pod \"35e40551-5ca0-42ef-8305-66fd69d36a9f\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.721342 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume" (OuterVolumeSpecName: "config-volume") pod "35e40551-5ca0-42ef-8305-66fd69d36a9f" (UID: "35e40551-5ca0-42ef-8305-66fd69d36a9f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.721489 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.725728 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7" (OuterVolumeSpecName: "kube-api-access-jgsr7") pod "35e40551-5ca0-42ef-8305-66fd69d36a9f" (UID: "35e40551-5ca0-42ef-8305-66fd69d36a9f"). InnerVolumeSpecName "kube-api-access-jgsr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.725744 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "35e40551-5ca0-42ef-8305-66fd69d36a9f" (UID: "35e40551-5ca0-42ef-8305-66fd69d36a9f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.823456 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgsr7\" (UniqueName: \"kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7\") on node \"crc\" DevicePath \"\"" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.823500 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:45:03 crc kubenswrapper[4824]: I0121 11:45:03.332883 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" event={"ID":"35e40551-5ca0-42ef-8305-66fd69d36a9f","Type":"ContainerDied","Data":"77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310"} Jan 21 11:45:03 crc kubenswrapper[4824]: I0121 11:45:03.332919 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:03 crc kubenswrapper[4824]: I0121 11:45:03.332932 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310" Jan 21 11:45:03 crc kubenswrapper[4824]: I0121 11:45:03.624426 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz"] Jan 21 11:45:03 crc kubenswrapper[4824]: I0121 11:45:03.630563 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz"] Jan 21 11:45:04 crc kubenswrapper[4824]: I0121 11:45:04.056722 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09bd0b9c-46fb-433f-bd30-014c69c90d9f" path="/var/lib/kubelet/pods/09bd0b9c-46fb-433f-bd30-014c69c90d9f/volumes" Jan 21 11:46:00 crc kubenswrapper[4824]: I0121 11:46:00.758616 4824 scope.go:117] "RemoveContainer" containerID="277011d8fed4b1333d53f5a4a084181f52dada1e6d85b31750d9bab9937b3678" Jan 21 11:46:46 crc kubenswrapper[4824]: I0121 11:46:46.065075 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:46:46 crc kubenswrapper[4824]: I0121 11:46:46.065468 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:46:47 crc kubenswrapper[4824]: I0121 11:46:47.974023 4824 generic.go:334] "Generic (PLEG): container finished" podID="49b2a136-e5c3-4741-80c6-4c545d4a82a0" containerID="6cc48beb53c571878821f1833c1a75164e55e0ffea7eb2292d2abd90321cb03f" exitCode=0 Jan 21 11:46:47 crc kubenswrapper[4824]: I0121 11:46:47.974112 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" event={"ID":"49b2a136-e5c3-4741-80c6-4c545d4a82a0","Type":"ContainerDied","Data":"6cc48beb53c571878821f1833c1a75164e55e0ffea7eb2292d2abd90321cb03f"} Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.297347 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420017 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420072 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgqg9\" (UniqueName: \"kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420150 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420246 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420269 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420322 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.424468 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.424564 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9" (OuterVolumeSpecName: "kube-api-access-sgqg9") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "kube-api-access-sgqg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.440567 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory" (OuterVolumeSpecName: "inventory") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.440877 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.441805 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.441889 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.442924 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.522332 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.522728 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.522797 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.522863 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.522921 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.523010 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.523069 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgqg9\" (UniqueName: \"kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.990056 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" event={"ID":"49b2a136-e5c3-4741-80c6-4c545d4a82a0","Type":"ContainerDied","Data":"b8c227627e16ce045a124bfbe36c3ecb09476a5cdfcf39963427bafa05e52cee"} Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.990290 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8c227627e16ce045a124bfbe36c3ecb09476a5cdfcf39963427bafa05e52cee" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.990115 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.032567 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:09 crc kubenswrapper[4824]: E0121 11:47:09.033295 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e40551-5ca0-42ef-8305-66fd69d36a9f" containerName="collect-profiles" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.033309 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e40551-5ca0-42ef-8305-66fd69d36a9f" containerName="collect-profiles" Jan 21 11:47:09 crc kubenswrapper[4824]: E0121 11:47:09.033325 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b2a136-e5c3-4741-80c6-4c545d4a82a0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.033331 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b2a136-e5c3-4741-80c6-4c545d4a82a0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.033498 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e40551-5ca0-42ef-8305-66fd69d36a9f" containerName="collect-profiles" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.033517 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="49b2a136-e5c3-4741-80c6-4c545d4a82a0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.034669 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.046642 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.147752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.147799 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.148043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spzf5\" (UniqueName: \"kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.249919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.249975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.250030 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spzf5\" (UniqueName: \"kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.250421 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.250709 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.266598 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spzf5\" (UniqueName: \"kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.355160 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.753836 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:09 crc kubenswrapper[4824]: W0121 11:47:09.755296 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ffd9d0f_1714_416f_84a9_4f3ab0305ee2.slice/crio-e039f89cd27aa709b422bb17b8c143529a0e6be5ab4250e2f360bbd333a60f34 WatchSource:0}: Error finding container e039f89cd27aa709b422bb17b8c143529a0e6be5ab4250e2f360bbd333a60f34: Status 404 returned error can't find the container with id e039f89cd27aa709b422bb17b8c143529a0e6be5ab4250e2f360bbd333a60f34 Jan 21 11:47:10 crc kubenswrapper[4824]: E0121 11:47:10.004531 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ffd9d0f_1714_416f_84a9_4f3ab0305ee2.slice/crio-conmon-ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2.scope\": RecentStats: unable to find data in memory cache]" Jan 21 11:47:10 crc kubenswrapper[4824]: I0121 11:47:10.105035 4824 generic.go:334] "Generic (PLEG): container finished" podID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerID="ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2" exitCode=0 Jan 21 11:47:10 crc kubenswrapper[4824]: I0121 11:47:10.105075 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerDied","Data":"ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2"} Jan 21 11:47:10 crc kubenswrapper[4824]: I0121 11:47:10.105101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerStarted","Data":"e039f89cd27aa709b422bb17b8c143529a0e6be5ab4250e2f360bbd333a60f34"} Jan 21 11:47:11 crc kubenswrapper[4824]: I0121 11:47:11.133594 4824 generic.go:334] "Generic (PLEG): container finished" podID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerID="0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42" exitCode=0 Jan 21 11:47:11 crc kubenswrapper[4824]: I0121 11:47:11.133635 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerDied","Data":"0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42"} Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.140782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerStarted","Data":"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0"} Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.166208 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kwnpj" podStartSLOduration=1.657175521 podStartE2EDuration="3.166195009s" podCreationTimestamp="2026-01-21 11:47:09 +0000 UTC" firstStartedPulling="2026-01-21 11:47:10.106950281 +0000 UTC m=+2172.399979572" lastFinishedPulling="2026-01-21 11:47:11.615969768 +0000 UTC m=+2173.908999060" observedRunningTime="2026-01-21 11:47:12.152665756 +0000 UTC m=+2174.445695048" watchObservedRunningTime="2026-01-21 11:47:12.166195009 +0000 UTC m=+2174.459224301" Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.817245 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.820967 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.840872 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.911936 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.912068 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.912113 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8cq8\" (UniqueName: \"kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.013616 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.013702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.013740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8cq8\" (UniqueName: \"kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.014114 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.014168 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.033646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8cq8\" (UniqueName: \"kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.135583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.551888 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:14 crc kubenswrapper[4824]: I0121 11:47:14.164768 4824 generic.go:334] "Generic (PLEG): container finished" podID="06d23466-53be-467a-85af-c4c342b5c86f" containerID="062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b" exitCode=0 Jan 21 11:47:14 crc kubenswrapper[4824]: I0121 11:47:14.164918 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerDied","Data":"062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b"} Jan 21 11:47:14 crc kubenswrapper[4824]: I0121 11:47:14.164999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerStarted","Data":"baa513cc43814d362a686ad77f5a0ae6ae18c0565e4cb439ccbea8853383114f"} Jan 21 11:47:15 crc kubenswrapper[4824]: I0121 11:47:15.173193 4824 generic.go:334] "Generic (PLEG): container finished" podID="06d23466-53be-467a-85af-c4c342b5c86f" containerID="f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8" exitCode=0 Jan 21 11:47:15 crc kubenswrapper[4824]: I0121 11:47:15.173426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerDied","Data":"f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8"} Jan 21 11:47:16 crc kubenswrapper[4824]: I0121 11:47:16.065760 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:47:16 crc kubenswrapper[4824]: I0121 11:47:16.066178 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:47:16 crc kubenswrapper[4824]: I0121 11:47:16.181730 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerStarted","Data":"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a"} Jan 21 11:47:16 crc kubenswrapper[4824]: I0121 11:47:16.194707 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kkvsf" podStartSLOduration=2.721735921 podStartE2EDuration="4.194693702s" podCreationTimestamp="2026-01-21 11:47:12 +0000 UTC" firstStartedPulling="2026-01-21 11:47:14.166241608 +0000 UTC m=+2176.459270900" lastFinishedPulling="2026-01-21 11:47:15.639199399 +0000 UTC m=+2177.932228681" observedRunningTime="2026-01-21 11:47:16.19405051 +0000 UTC m=+2178.487079802" watchObservedRunningTime="2026-01-21 11:47:16.194693702 +0000 UTC m=+2178.487722994" Jan 21 11:47:19 crc kubenswrapper[4824]: I0121 11:47:19.355713 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:19 crc kubenswrapper[4824]: I0121 11:47:19.356049 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:19 crc kubenswrapper[4824]: I0121 11:47:19.389968 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:20 crc kubenswrapper[4824]: I0121 11:47:20.241184 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:20 crc kubenswrapper[4824]: I0121 11:47:20.605560 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.218566 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kwnpj" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="registry-server" containerID="cri-o://27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0" gracePeriod=2 Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.567357 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.686613 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content\") pod \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.686898 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities\") pod \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.686935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spzf5\" (UniqueName: \"kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5\") pod \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.691781 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities" (OuterVolumeSpecName: "utilities") pod "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" (UID: "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.704361 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" (UID: "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.710560 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5" (OuterVolumeSpecName: "kube-api-access-spzf5") pod "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" (UID: "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2"). InnerVolumeSpecName "kube-api-access-spzf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.788731 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.788773 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spzf5\" (UniqueName: \"kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.788783 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.135850 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.136546 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.169528 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.225524 4824 generic.go:334] "Generic (PLEG): container finished" podID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerID="27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0" exitCode=0 Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.225589 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.225606 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerDied","Data":"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0"} Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.225633 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerDied","Data":"e039f89cd27aa709b422bb17b8c143529a0e6be5ab4250e2f360bbd333a60f34"} Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.225648 4824 scope.go:117] "RemoveContainer" containerID="27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.251085 4824 scope.go:117] "RemoveContainer" containerID="0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.252901 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.260365 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.265650 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.276797 4824 scope.go:117] "RemoveContainer" containerID="ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.303909 4824 scope.go:117] "RemoveContainer" containerID="27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0" Jan 21 11:47:23 crc kubenswrapper[4824]: E0121 11:47:23.304434 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0\": container with ID starting with 27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0 not found: ID does not exist" containerID="27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.304460 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0"} err="failed to get container status \"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0\": rpc error: code = NotFound desc = could not find container \"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0\": container with ID starting with 27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0 not found: ID does not exist" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.304479 4824 scope.go:117] "RemoveContainer" containerID="0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42" Jan 21 11:47:23 crc kubenswrapper[4824]: E0121 11:47:23.304723 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42\": container with ID starting with 0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42 not found: ID does not exist" containerID="0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.304742 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42"} err="failed to get container status \"0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42\": rpc error: code = NotFound desc = could not find container \"0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42\": container with ID starting with 0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42 not found: ID does not exist" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.304755 4824 scope.go:117] "RemoveContainer" containerID="ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2" Jan 21 11:47:23 crc kubenswrapper[4824]: E0121 11:47:23.305065 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2\": container with ID starting with ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2 not found: ID does not exist" containerID="ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.305087 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2"} err="failed to get container status \"ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2\": rpc error: code = NotFound desc = could not find container \"ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2\": container with ID starting with ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2 not found: ID does not exist" Jan 21 11:47:24 crc kubenswrapper[4824]: I0121 11:47:24.058336 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" path="/var/lib/kubelet/pods/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2/volumes" Jan 21 11:47:24 crc kubenswrapper[4824]: I0121 11:47:24.608237 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.244086 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kkvsf" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="registry-server" containerID="cri-o://105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a" gracePeriod=2 Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.596592 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.758214 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8cq8\" (UniqueName: \"kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8\") pod \"06d23466-53be-467a-85af-c4c342b5c86f\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.758341 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content\") pod \"06d23466-53be-467a-85af-c4c342b5c86f\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.758446 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities\") pod \"06d23466-53be-467a-85af-c4c342b5c86f\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.758997 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities" (OuterVolumeSpecName: "utilities") pod "06d23466-53be-467a-85af-c4c342b5c86f" (UID: "06d23466-53be-467a-85af-c4c342b5c86f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.763153 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8" (OuterVolumeSpecName: "kube-api-access-d8cq8") pod "06d23466-53be-467a-85af-c4c342b5c86f" (UID: "06d23466-53be-467a-85af-c4c342b5c86f"). InnerVolumeSpecName "kube-api-access-d8cq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.794527 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06d23466-53be-467a-85af-c4c342b5c86f" (UID: "06d23466-53be-467a-85af-c4c342b5c86f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.860380 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.860404 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.860414 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8cq8\" (UniqueName: \"kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.252618 4824 generic.go:334] "Generic (PLEG): container finished" podID="06d23466-53be-467a-85af-c4c342b5c86f" containerID="105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a" exitCode=0 Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.252658 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.252675 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerDied","Data":"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a"} Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.252915 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerDied","Data":"baa513cc43814d362a686ad77f5a0ae6ae18c0565e4cb439ccbea8853383114f"} Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.252932 4824 scope.go:117] "RemoveContainer" containerID="105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.267388 4824 scope.go:117] "RemoveContainer" containerID="f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.278614 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.284843 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.296320 4824 scope.go:117] "RemoveContainer" containerID="062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.316257 4824 scope.go:117] "RemoveContainer" containerID="105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a" Jan 21 11:47:27 crc kubenswrapper[4824]: E0121 11:47:27.316601 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a\": container with ID starting with 105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a not found: ID does not exist" containerID="105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.316637 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a"} err="failed to get container status \"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a\": rpc error: code = NotFound desc = could not find container \"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a\": container with ID starting with 105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a not found: ID does not exist" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.316658 4824 scope.go:117] "RemoveContainer" containerID="f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8" Jan 21 11:47:27 crc kubenswrapper[4824]: E0121 11:47:27.316988 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8\": container with ID starting with f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8 not found: ID does not exist" containerID="f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.317104 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8"} err="failed to get container status \"f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8\": rpc error: code = NotFound desc = could not find container \"f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8\": container with ID starting with f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8 not found: ID does not exist" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.317200 4824 scope.go:117] "RemoveContainer" containerID="062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b" Jan 21 11:47:27 crc kubenswrapper[4824]: E0121 11:47:27.317552 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b\": container with ID starting with 062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b not found: ID does not exist" containerID="062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.317577 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b"} err="failed to get container status \"062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b\": rpc error: code = NotFound desc = could not find container \"062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b\": container with ID starting with 062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b not found: ID does not exist" Jan 21 11:47:28 crc kubenswrapper[4824]: I0121 11:47:28.056299 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06d23466-53be-467a-85af-c4c342b5c86f" path="/var/lib/kubelet/pods/06d23466-53be-467a-85af-c4c342b5c86f/volumes" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103059 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103575 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="extract-content" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103588 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="extract-content" Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103604 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="extract-content" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103610 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="extract-content" Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103625 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103631 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103639 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="extract-utilities" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103645 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="extract-utilities" Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103668 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="extract-utilities" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103673 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="extract-utilities" Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103682 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103687 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103848 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103868 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.104421 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.105790 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qs54z" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.106017 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.106035 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.106055 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.109575 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.263743 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr6pd\" (UniqueName: \"kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.263816 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264096 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264165 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264229 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264303 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.365975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366061 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366123 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366148 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366344 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366551 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.367197 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr6pd\" (UniqueName: \"kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.367265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.367309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.367839 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.371821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.372033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.372150 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.383921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr6pd\" (UniqueName: \"kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.385945 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.426556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.846055 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Jan 21 11:47:34 crc kubenswrapper[4824]: I0121 11:47:34.296532 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"58d30951-ee56-4217-a614-a04951369c8b","Type":"ContainerStarted","Data":"5bc5adacec79286d2293a6c562788392bcdca8f811b129442be3e2105c5568bf"} Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.065544 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.065905 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.065940 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.066375 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.066422 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" gracePeriod=600 Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.383245 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" exitCode=0 Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.383290 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788"} Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.383336 4824 scope.go:117] "RemoveContainer" containerID="1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.241843 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.321064 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.321435 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dr6pd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(58d30951-ee56-4217-a614-a04951369c8b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.323100 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="58d30951-ee56-4217-a614-a04951369c8b" Jan 21 11:47:54 crc kubenswrapper[4824]: I0121 11:47:54.444025 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.444362 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.444641 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="58d30951-ee56-4217-a614-a04951369c8b" Jan 21 11:48:08 crc kubenswrapper[4824]: I0121 11:48:08.053365 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:48:08 crc kubenswrapper[4824]: E0121 11:48:08.054056 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:48:09 crc kubenswrapper[4824]: I0121 11:48:09.569255 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Jan 21 11:48:10 crc kubenswrapper[4824]: I0121 11:48:10.553519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"58d30951-ee56-4217-a614-a04951369c8b","Type":"ContainerStarted","Data":"f56ac3b5212e3537571012922916244691168864ad9d36d1f4cde9480e97ed7a"} Jan 21 11:48:10 crc kubenswrapper[4824]: I0121 11:48:10.575737 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=2.8575966729999998 podStartE2EDuration="38.575717158s" podCreationTimestamp="2026-01-21 11:47:32 +0000 UTC" firstStartedPulling="2026-01-21 11:47:33.848737983 +0000 UTC m=+2196.141767275" lastFinishedPulling="2026-01-21 11:48:09.566858467 +0000 UTC m=+2231.859887760" observedRunningTime="2026-01-21 11:48:10.568096411 +0000 UTC m=+2232.861125713" watchObservedRunningTime="2026-01-21 11:48:10.575717158 +0000 UTC m=+2232.868746449" Jan 21 11:48:23 crc kubenswrapper[4824]: I0121 11:48:23.049603 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:48:23 crc kubenswrapper[4824]: E0121 11:48:23.050192 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:48:38 crc kubenswrapper[4824]: I0121 11:48:38.049999 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:48:38 crc kubenswrapper[4824]: E0121 11:48:38.051009 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:48:51 crc kubenswrapper[4824]: I0121 11:48:51.049873 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:48:51 crc kubenswrapper[4824]: E0121 11:48:51.051468 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:49:05 crc kubenswrapper[4824]: I0121 11:49:05.050016 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:49:05 crc kubenswrapper[4824]: E0121 11:49:05.050731 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:49:18 crc kubenswrapper[4824]: I0121 11:49:18.053759 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:49:18 crc kubenswrapper[4824]: E0121 11:49:18.054310 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:49:32 crc kubenswrapper[4824]: I0121 11:49:32.053014 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:49:32 crc kubenswrapper[4824]: E0121 11:49:32.053768 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:49:43 crc kubenswrapper[4824]: I0121 11:49:43.049785 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:49:43 crc kubenswrapper[4824]: E0121 11:49:43.050446 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:49:57 crc kubenswrapper[4824]: I0121 11:49:57.049223 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:49:57 crc kubenswrapper[4824]: E0121 11:49:57.049801 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:50:12 crc kubenswrapper[4824]: I0121 11:50:12.051546 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:50:12 crc kubenswrapper[4824]: E0121 11:50:12.052049 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:50:26 crc kubenswrapper[4824]: I0121 11:50:26.049548 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:50:26 crc kubenswrapper[4824]: E0121 11:50:26.050258 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:50:38 crc kubenswrapper[4824]: I0121 11:50:38.053760 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:50:38 crc kubenswrapper[4824]: E0121 11:50:38.054549 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:50:49 crc kubenswrapper[4824]: I0121 11:50:49.049268 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:50:49 crc kubenswrapper[4824]: E0121 11:50:49.049819 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:51:02 crc kubenswrapper[4824]: I0121 11:51:02.050508 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:51:02 crc kubenswrapper[4824]: E0121 11:51:02.051503 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:51:15 crc kubenswrapper[4824]: I0121 11:51:15.049793 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:51:15 crc kubenswrapper[4824]: E0121 11:51:15.050401 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:51:28 crc kubenswrapper[4824]: I0121 11:51:28.053787 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:51:28 crc kubenswrapper[4824]: E0121 11:51:28.054512 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:51:40 crc kubenswrapper[4824]: I0121 11:51:40.049410 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:51:40 crc kubenswrapper[4824]: E0121 11:51:40.050079 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:51:52 crc kubenswrapper[4824]: I0121 11:51:52.049780 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:51:52 crc kubenswrapper[4824]: E0121 11:51:52.050515 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:52:06 crc kubenswrapper[4824]: I0121 11:52:06.049735 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:52:06 crc kubenswrapper[4824]: E0121 11:52:06.050643 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:52:19 crc kubenswrapper[4824]: I0121 11:52:19.049638 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:52:19 crc kubenswrapper[4824]: E0121 11:52:19.050211 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:52:33 crc kubenswrapper[4824]: I0121 11:52:33.050144 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:52:33 crc kubenswrapper[4824]: E0121 11:52:33.050686 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.618342 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.620093 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.639412 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.704323 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.704492 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swv92\" (UniqueName: \"kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.704525 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.805664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swv92\" (UniqueName: \"kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.805703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.805877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.806284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.806340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.824913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swv92\" (UniqueName: \"kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.937309 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:36 crc kubenswrapper[4824]: I0121 11:52:36.492817 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:37 crc kubenswrapper[4824]: I0121 11:52:37.284111 4824 generic.go:334] "Generic (PLEG): container finished" podID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerID="5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce" exitCode=0 Jan 21 11:52:37 crc kubenswrapper[4824]: I0121 11:52:37.284167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerDied","Data":"5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce"} Jan 21 11:52:37 crc kubenswrapper[4824]: I0121 11:52:37.284437 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerStarted","Data":"2f09f6b203e5c39ff281636d181c83fa7016bbf854baf158e8e90be501cf2e21"} Jan 21 11:52:37 crc kubenswrapper[4824]: I0121 11:52:37.285879 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.293990 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerStarted","Data":"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e"} Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.817321 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.819113 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.839801 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.861168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tntsc\" (UniqueName: \"kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.861221 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.861266 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.963399 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tntsc\" (UniqueName: \"kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.963463 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.963511 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.964069 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.964199 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.985654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tntsc\" (UniqueName: \"kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:39 crc kubenswrapper[4824]: I0121 11:52:39.141089 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:39 crc kubenswrapper[4824]: I0121 11:52:39.305643 4824 generic.go:334] "Generic (PLEG): container finished" podID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerID="6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e" exitCode=0 Jan 21 11:52:39 crc kubenswrapper[4824]: I0121 11:52:39.305748 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerDied","Data":"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e"} Jan 21 11:52:39 crc kubenswrapper[4824]: I0121 11:52:39.538266 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:52:39 crc kubenswrapper[4824]: W0121 11:52:39.543769 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf557c9a4_cd4b_467b_8d42_43c9f6e2d6f0.slice/crio-749644413b97fec9cd23377ee2def7d6bcb9e73627dcf6518eaa3d3461f70acc WatchSource:0}: Error finding container 749644413b97fec9cd23377ee2def7d6bcb9e73627dcf6518eaa3d3461f70acc: Status 404 returned error can't find the container with id 749644413b97fec9cd23377ee2def7d6bcb9e73627dcf6518eaa3d3461f70acc Jan 21 11:52:40 crc kubenswrapper[4824]: I0121 11:52:40.313831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerStarted","Data":"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97"} Jan 21 11:52:40 crc kubenswrapper[4824]: I0121 11:52:40.315485 4824 generic.go:334] "Generic (PLEG): container finished" podID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerID="578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9" exitCode=0 Jan 21 11:52:40 crc kubenswrapper[4824]: I0121 11:52:40.315522 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerDied","Data":"578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9"} Jan 21 11:52:40 crc kubenswrapper[4824]: I0121 11:52:40.315567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerStarted","Data":"749644413b97fec9cd23377ee2def7d6bcb9e73627dcf6518eaa3d3461f70acc"} Jan 21 11:52:40 crc kubenswrapper[4824]: I0121 11:52:40.335450 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-94jm9" podStartSLOduration=2.717146636 podStartE2EDuration="5.335437768s" podCreationTimestamp="2026-01-21 11:52:35 +0000 UTC" firstStartedPulling="2026-01-21 11:52:37.285679871 +0000 UTC m=+2499.578709163" lastFinishedPulling="2026-01-21 11:52:39.903971003 +0000 UTC m=+2502.197000295" observedRunningTime="2026-01-21 11:52:40.329934012 +0000 UTC m=+2502.622963305" watchObservedRunningTime="2026-01-21 11:52:40.335437768 +0000 UTC m=+2502.628467060" Jan 21 11:52:41 crc kubenswrapper[4824]: I0121 11:52:41.322823 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerStarted","Data":"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3"} Jan 21 11:52:44 crc kubenswrapper[4824]: I0121 11:52:44.050378 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:52:44 crc kubenswrapper[4824]: E0121 11:52:44.050584 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:52:44 crc kubenswrapper[4824]: I0121 11:52:44.341833 4824 generic.go:334] "Generic (PLEG): container finished" podID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerID="068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3" exitCode=0 Jan 21 11:52:44 crc kubenswrapper[4824]: I0121 11:52:44.341923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerDied","Data":"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3"} Jan 21 11:52:45 crc kubenswrapper[4824]: I0121 11:52:45.357027 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerStarted","Data":"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b"} Jan 21 11:52:45 crc kubenswrapper[4824]: I0121 11:52:45.377785 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f9255" podStartSLOduration=2.867260868 podStartE2EDuration="7.377768786s" podCreationTimestamp="2026-01-21 11:52:38 +0000 UTC" firstStartedPulling="2026-01-21 11:52:40.317144439 +0000 UTC m=+2502.610173731" lastFinishedPulling="2026-01-21 11:52:44.827652357 +0000 UTC m=+2507.120681649" observedRunningTime="2026-01-21 11:52:45.371680447 +0000 UTC m=+2507.664709739" watchObservedRunningTime="2026-01-21 11:52:45.377768786 +0000 UTC m=+2507.670798067" Jan 21 11:52:45 crc kubenswrapper[4824]: I0121 11:52:45.937674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:45 crc kubenswrapper[4824]: I0121 11:52:45.937741 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:45 crc kubenswrapper[4824]: I0121 11:52:45.972225 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:46 crc kubenswrapper[4824]: I0121 11:52:46.395281 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:47 crc kubenswrapper[4824]: I0121 11:52:47.007399 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:48 crc kubenswrapper[4824]: I0121 11:52:48.374165 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-94jm9" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="registry-server" containerID="cri-o://8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97" gracePeriod=2 Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.141915 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.142068 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.309681 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.350824 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content\") pod \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.350875 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities\") pod \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.351028 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swv92\" (UniqueName: \"kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92\") pod \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.351403 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities" (OuterVolumeSpecName: "utilities") pod "17ee7b33-8d12-49da-b45b-5b8d75e53a08" (UID: "17ee7b33-8d12-49da-b45b-5b8d75e53a08"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.351783 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.355634 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92" (OuterVolumeSpecName: "kube-api-access-swv92") pod "17ee7b33-8d12-49da-b45b-5b8d75e53a08" (UID: "17ee7b33-8d12-49da-b45b-5b8d75e53a08"). InnerVolumeSpecName "kube-api-access-swv92". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.382199 4824 generic.go:334] "Generic (PLEG): container finished" podID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerID="8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97" exitCode=0 Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.382261 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.382264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerDied","Data":"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97"} Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.382931 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerDied","Data":"2f09f6b203e5c39ff281636d181c83fa7016bbf854baf158e8e90be501cf2e21"} Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.382952 4824 scope.go:117] "RemoveContainer" containerID="8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.386494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17ee7b33-8d12-49da-b45b-5b8d75e53a08" (UID: "17ee7b33-8d12-49da-b45b-5b8d75e53a08"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.399493 4824 scope.go:117] "RemoveContainer" containerID="6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.416985 4824 scope.go:117] "RemoveContainer" containerID="5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.446474 4824 scope.go:117] "RemoveContainer" containerID="8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97" Jan 21 11:52:49 crc kubenswrapper[4824]: E0121 11:52:49.446763 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97\": container with ID starting with 8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97 not found: ID does not exist" containerID="8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.446792 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97"} err="failed to get container status \"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97\": rpc error: code = NotFound desc = could not find container \"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97\": container with ID starting with 8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97 not found: ID does not exist" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.446811 4824 scope.go:117] "RemoveContainer" containerID="6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e" Jan 21 11:52:49 crc kubenswrapper[4824]: E0121 11:52:49.447063 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e\": container with ID starting with 6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e not found: ID does not exist" containerID="6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.447099 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e"} err="failed to get container status \"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e\": rpc error: code = NotFound desc = could not find container \"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e\": container with ID starting with 6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e not found: ID does not exist" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.447121 4824 scope.go:117] "RemoveContainer" containerID="5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce" Jan 21 11:52:49 crc kubenswrapper[4824]: E0121 11:52:49.447403 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce\": container with ID starting with 5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce not found: ID does not exist" containerID="5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.447426 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce"} err="failed to get container status \"5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce\": rpc error: code = NotFound desc = could not find container \"5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce\": container with ID starting with 5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce not found: ID does not exist" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.453612 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swv92\" (UniqueName: \"kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92\") on node \"crc\" DevicePath \"\"" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.453640 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.708022 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.713971 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:50 crc kubenswrapper[4824]: I0121 11:52:50.056943 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" path="/var/lib/kubelet/pods/17ee7b33-8d12-49da-b45b-5b8d75e53a08/volumes" Jan 21 11:52:50 crc kubenswrapper[4824]: I0121 11:52:50.177259 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f9255" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="registry-server" probeResult="failure" output=< Jan 21 11:52:50 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Jan 21 11:52:50 crc kubenswrapper[4824]: > Jan 21 11:52:58 crc kubenswrapper[4824]: I0121 11:52:58.054311 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:52:58 crc kubenswrapper[4824]: I0121 11:52:58.445942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501"} Jan 21 11:52:59 crc kubenswrapper[4824]: I0121 11:52:59.174211 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:59 crc kubenswrapper[4824]: I0121 11:52:59.206558 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:59 crc kubenswrapper[4824]: I0121 11:52:59.403129 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.457272 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f9255" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="registry-server" containerID="cri-o://e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b" gracePeriod=2 Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.835098 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.948819 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content\") pod \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.948876 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tntsc\" (UniqueName: \"kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc\") pod \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.948995 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities\") pod \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.949579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities" (OuterVolumeSpecName: "utilities") pod "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" (UID: "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.955323 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc" (OuterVolumeSpecName: "kube-api-access-tntsc") pod "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" (UID: "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0"). InnerVolumeSpecName "kube-api-access-tntsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.044543 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" (UID: "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.050937 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.050981 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.050992 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tntsc\" (UniqueName: \"kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc\") on node \"crc\" DevicePath \"\"" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.466342 4824 generic.go:334] "Generic (PLEG): container finished" podID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerID="e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b" exitCode=0 Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.466439 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.466463 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerDied","Data":"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b"} Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.467420 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerDied","Data":"749644413b97fec9cd23377ee2def7d6bcb9e73627dcf6518eaa3d3461f70acc"} Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.467463 4824 scope.go:117] "RemoveContainer" containerID="e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.488048 4824 scope.go:117] "RemoveContainer" containerID="068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.495208 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.500193 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.520038 4824 scope.go:117] "RemoveContainer" containerID="578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.535609 4824 scope.go:117] "RemoveContainer" containerID="e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b" Jan 21 11:53:01 crc kubenswrapper[4824]: E0121 11:53:01.535973 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b\": container with ID starting with e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b not found: ID does not exist" containerID="e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.536009 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b"} err="failed to get container status \"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b\": rpc error: code = NotFound desc = could not find container \"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b\": container with ID starting with e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b not found: ID does not exist" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.536033 4824 scope.go:117] "RemoveContainer" containerID="068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3" Jan 21 11:53:01 crc kubenswrapper[4824]: E0121 11:53:01.536311 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3\": container with ID starting with 068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3 not found: ID does not exist" containerID="068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.536338 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3"} err="failed to get container status \"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3\": rpc error: code = NotFound desc = could not find container \"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3\": container with ID starting with 068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3 not found: ID does not exist" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.536358 4824 scope.go:117] "RemoveContainer" containerID="578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9" Jan 21 11:53:01 crc kubenswrapper[4824]: E0121 11:53:01.536597 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9\": container with ID starting with 578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9 not found: ID does not exist" containerID="578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.536630 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9"} err="failed to get container status \"578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9\": rpc error: code = NotFound desc = could not find container \"578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9\": container with ID starting with 578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9 not found: ID does not exist" Jan 21 11:53:02 crc kubenswrapper[4824]: I0121 11:53:02.057385 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" path="/var/lib/kubelet/pods/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0/volumes" Jan 21 11:55:16 crc kubenswrapper[4824]: I0121 11:55:16.065233 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:55:16 crc kubenswrapper[4824]: I0121 11:55:16.065654 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:55:46 crc kubenswrapper[4824]: I0121 11:55:46.065145 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:55:46 crc kubenswrapper[4824]: I0121 11:55:46.065514 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.065216 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.065612 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.065644 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.066018 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.066057 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501" gracePeriod=600 Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.694012 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501" exitCode=0 Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.694094 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501"} Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.694366 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a"} Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.694386 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.889321 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890002 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="extract-utilities" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890015 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="extract-utilities" Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890025 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="extract-content" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890031 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="extract-content" Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890041 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890046 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890077 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="extract-utilities" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890083 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="extract-utilities" Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890095 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="extract-content" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890101 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="extract-content" Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890116 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890122 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890306 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890327 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.891414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.897281 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.085865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.085913 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cfdg\" (UniqueName: \"kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.086153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.188331 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.188377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cfdg\" (UniqueName: \"kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.188591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.188791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.188990 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.205552 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cfdg\" (UniqueName: \"kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.205853 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.622745 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:45 crc kubenswrapper[4824]: I0121 11:57:45.258188 4824 generic.go:334] "Generic (PLEG): container finished" podID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerID="49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982" exitCode=0 Jan 21 11:57:45 crc kubenswrapper[4824]: I0121 11:57:45.258250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerDied","Data":"49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982"} Jan 21 11:57:45 crc kubenswrapper[4824]: I0121 11:57:45.258424 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerStarted","Data":"277001b0b5c366eff7d23d99e1d7e02281584f2e7f61c7ed99594590aa45edfa"} Jan 21 11:57:45 crc kubenswrapper[4824]: I0121 11:57:45.260750 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.091593 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.093749 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.098811 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.226452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.228670 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.228833 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsv99\" (UniqueName: \"kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.266974 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerStarted","Data":"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50"} Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.331641 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.331741 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsv99\" (UniqueName: \"kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.331822 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.332012 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.332305 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.352699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsv99\" (UniqueName: \"kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.413357 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.821661 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:57:47 crc kubenswrapper[4824]: I0121 11:57:47.275111 4824 generic.go:334] "Generic (PLEG): container finished" podID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerID="4348e65f757892dad73a8da976eba5527fd6deb6ba5e99e3204b12a7526894f3" exitCode=0 Jan 21 11:57:47 crc kubenswrapper[4824]: I0121 11:57:47.275185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerDied","Data":"4348e65f757892dad73a8da976eba5527fd6deb6ba5e99e3204b12a7526894f3"} Jan 21 11:57:47 crc kubenswrapper[4824]: I0121 11:57:47.275216 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerStarted","Data":"6aa35848e24a34105b74ff6810b0b5a65321efe98eb936afe9a804fe7f741e36"} Jan 21 11:57:47 crc kubenswrapper[4824]: I0121 11:57:47.276882 4824 generic.go:334] "Generic (PLEG): container finished" podID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerID="638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50" exitCode=0 Jan 21 11:57:47 crc kubenswrapper[4824]: I0121 11:57:47.276924 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerDied","Data":"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50"} Jan 21 11:57:48 crc kubenswrapper[4824]: I0121 11:57:48.289556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerStarted","Data":"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef"} Jan 21 11:57:48 crc kubenswrapper[4824]: I0121 11:57:48.292253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerStarted","Data":"1395c6331c55dda72692c82cfb7f405a6f14ad9c4c33c90a877304f74be6c19d"} Jan 21 11:57:48 crc kubenswrapper[4824]: I0121 11:57:48.316133 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f9xtx" podStartSLOduration=2.845292348 podStartE2EDuration="5.316117569s" podCreationTimestamp="2026-01-21 11:57:43 +0000 UTC" firstStartedPulling="2026-01-21 11:57:45.260524009 +0000 UTC m=+2807.553553301" lastFinishedPulling="2026-01-21 11:57:47.73134923 +0000 UTC m=+2810.024378522" observedRunningTime="2026-01-21 11:57:48.308101306 +0000 UTC m=+2810.601130598" watchObservedRunningTime="2026-01-21 11:57:48.316117569 +0000 UTC m=+2810.609146850" Jan 21 11:57:49 crc kubenswrapper[4824]: I0121 11:57:49.300626 4824 generic.go:334] "Generic (PLEG): container finished" podID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerID="1395c6331c55dda72692c82cfb7f405a6f14ad9c4c33c90a877304f74be6c19d" exitCode=0 Jan 21 11:57:49 crc kubenswrapper[4824]: I0121 11:57:49.300713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerDied","Data":"1395c6331c55dda72692c82cfb7f405a6f14ad9c4c33c90a877304f74be6c19d"} Jan 21 11:57:50 crc kubenswrapper[4824]: I0121 11:57:50.311151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerStarted","Data":"56a664df70a0da76de98566146b77a6f9c3a743b2a196b8ba3f9c56f6ec5dd11"} Jan 21 11:57:50 crc kubenswrapper[4824]: I0121 11:57:50.333859 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dwwjr" podStartSLOduration=1.835427191 podStartE2EDuration="4.333843129s" podCreationTimestamp="2026-01-21 11:57:46 +0000 UTC" firstStartedPulling="2026-01-21 11:57:47.276396156 +0000 UTC m=+2809.569425447" lastFinishedPulling="2026-01-21 11:57:49.774812093 +0000 UTC m=+2812.067841385" observedRunningTime="2026-01-21 11:57:50.326997343 +0000 UTC m=+2812.620026635" watchObservedRunningTime="2026-01-21 11:57:50.333843129 +0000 UTC m=+2812.626872411" Jan 21 11:57:54 crc kubenswrapper[4824]: I0121 11:57:54.207094 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:54 crc kubenswrapper[4824]: I0121 11:57:54.207493 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:54 crc kubenswrapper[4824]: I0121 11:57:54.238211 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:54 crc kubenswrapper[4824]: I0121 11:57:54.367183 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:54 crc kubenswrapper[4824]: I0121 11:57:54.881753 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.348408 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f9xtx" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="registry-server" containerID="cri-o://2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef" gracePeriod=2 Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.413460 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.413518 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.450445 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.735361 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.830998 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cfdg\" (UniqueName: \"kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg\") pod \"95ce4f3d-0b58-42b2-9930-23952f0be687\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.831323 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities\") pod \"95ce4f3d-0b58-42b2-9930-23952f0be687\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.831419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content\") pod \"95ce4f3d-0b58-42b2-9930-23952f0be687\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.831995 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities" (OuterVolumeSpecName: "utilities") pod "95ce4f3d-0b58-42b2-9930-23952f0be687" (UID: "95ce4f3d-0b58-42b2-9930-23952f0be687"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.832331 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.844676 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg" (OuterVolumeSpecName: "kube-api-access-9cfdg") pod "95ce4f3d-0b58-42b2-9930-23952f0be687" (UID: "95ce4f3d-0b58-42b2-9930-23952f0be687"). InnerVolumeSpecName "kube-api-access-9cfdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.869394 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95ce4f3d-0b58-42b2-9930-23952f0be687" (UID: "95ce4f3d-0b58-42b2-9930-23952f0be687"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.934529 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cfdg\" (UniqueName: \"kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg\") on node \"crc\" DevicePath \"\"" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.934690 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.372366 4824 generic.go:334] "Generic (PLEG): container finished" podID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerID="2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef" exitCode=0 Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.372488 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.372529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerDied","Data":"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef"} Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.372571 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerDied","Data":"277001b0b5c366eff7d23d99e1d7e02281584f2e7f61c7ed99594590aa45edfa"} Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.372588 4824 scope.go:117] "RemoveContainer" containerID="2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.391420 4824 scope.go:117] "RemoveContainer" containerID="638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.400582 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.407568 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.409499 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.424835 4824 scope.go:117] "RemoveContainer" containerID="49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.440549 4824 scope.go:117] "RemoveContainer" containerID="2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef" Jan 21 11:57:57 crc kubenswrapper[4824]: E0121 11:57:57.440931 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef\": container with ID starting with 2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef not found: ID does not exist" containerID="2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.440996 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef"} err="failed to get container status \"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef\": rpc error: code = NotFound desc = could not find container \"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef\": container with ID starting with 2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef not found: ID does not exist" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.441020 4824 scope.go:117] "RemoveContainer" containerID="638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50" Jan 21 11:57:57 crc kubenswrapper[4824]: E0121 11:57:57.441307 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50\": container with ID starting with 638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50 not found: ID does not exist" containerID="638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.441340 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50"} err="failed to get container status \"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50\": rpc error: code = NotFound desc = could not find container \"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50\": container with ID starting with 638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50 not found: ID does not exist" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.441382 4824 scope.go:117] "RemoveContainer" containerID="49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982" Jan 21 11:57:57 crc kubenswrapper[4824]: E0121 11:57:57.441763 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982\": container with ID starting with 49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982 not found: ID does not exist" containerID="49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.441785 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982"} err="failed to get container status \"49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982\": rpc error: code = NotFound desc = could not find container \"49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982\": container with ID starting with 49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982 not found: ID does not exist" Jan 21 11:57:58 crc kubenswrapper[4824]: I0121 11:57:58.058234 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" path="/var/lib/kubelet/pods/95ce4f3d-0b58-42b2-9930-23952f0be687/volumes" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.281934 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.282499 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dwwjr" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="registry-server" containerID="cri-o://56a664df70a0da76de98566146b77a6f9c3a743b2a196b8ba3f9c56f6ec5dd11" gracePeriod=2 Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.415459 4824 generic.go:334] "Generic (PLEG): container finished" podID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerID="56a664df70a0da76de98566146b77a6f9c3a743b2a196b8ba3f9c56f6ec5dd11" exitCode=0 Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.415700 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerDied","Data":"56a664df70a0da76de98566146b77a6f9c3a743b2a196b8ba3f9c56f6ec5dd11"} Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.689479 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.722978 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content\") pod \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.723032 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities\") pod \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.723231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsv99\" (UniqueName: \"kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99\") pod \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.723973 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities" (OuterVolumeSpecName: "utilities") pod "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" (UID: "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.728171 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99" (OuterVolumeSpecName: "kube-api-access-gsv99") pod "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" (UID: "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe"). InnerVolumeSpecName "kube-api-access-gsv99". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.739222 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" (UID: "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.826163 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsv99\" (UniqueName: \"kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99\") on node \"crc\" DevicePath \"\"" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.826435 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.826446 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.424814 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerDied","Data":"6aa35848e24a34105b74ff6810b0b5a65321efe98eb936afe9a804fe7f741e36"} Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.424863 4824 scope.go:117] "RemoveContainer" containerID="56a664df70a0da76de98566146b77a6f9c3a743b2a196b8ba3f9c56f6ec5dd11" Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.424865 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.442087 4824 scope.go:117] "RemoveContainer" containerID="1395c6331c55dda72692c82cfb7f405a6f14ad9c4c33c90a877304f74be6c19d" Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.443396 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.451285 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.457663 4824 scope.go:117] "RemoveContainer" containerID="4348e65f757892dad73a8da976eba5527fd6deb6ba5e99e3204b12a7526894f3" Jan 21 11:58:04 crc kubenswrapper[4824]: I0121 11:58:04.056829 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" path="/var/lib/kubelet/pods/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe/volumes" Jan 21 11:58:16 crc kubenswrapper[4824]: I0121 11:58:16.064940 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:58:16 crc kubenswrapper[4824]: I0121 11:58:16.065635 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:58:46 crc kubenswrapper[4824]: I0121 11:58:46.065040 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:58:46 crc kubenswrapper[4824]: I0121 11:58:46.065513 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.064845 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.065249 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.065289 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.065691 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.065743 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" gracePeriod=600 Jan 21 11:59:16 crc kubenswrapper[4824]: E0121 11:59:16.180343 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.910739 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" exitCode=0 Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.910782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a"} Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.910823 4824 scope.go:117] "RemoveContainer" containerID="54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.911450 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 11:59:16 crc kubenswrapper[4824]: E0121 11:59:16.911891 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:59:28 crc kubenswrapper[4824]: I0121 11:59:28.054513 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 11:59:28 crc kubenswrapper[4824]: E0121 11:59:28.055251 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:59:40 crc kubenswrapper[4824]: I0121 11:59:40.049230 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 11:59:40 crc kubenswrapper[4824]: E0121 11:59:40.050523 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:59:53 crc kubenswrapper[4824]: I0121 11:59:53.133829 4824 generic.go:334] "Generic (PLEG): container finished" podID="58d30951-ee56-4217-a614-a04951369c8b" containerID="f56ac3b5212e3537571012922916244691168864ad9d36d1f4cde9480e97ed7a" exitCode=0 Jan 21 11:59:53 crc kubenswrapper[4824]: I0121 11:59:53.133925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"58d30951-ee56-4217-a614-a04951369c8b","Type":"ContainerDied","Data":"f56ac3b5212e3537571012922916244691168864ad9d36d1f4cde9480e97ed7a"} Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.049810 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 11:59:54 crc kubenswrapper[4824]: E0121 11:59:54.050032 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.563024 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.604887 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.604935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.604998 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605113 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dr6pd\" (UniqueName: \"kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605711 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605740 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605806 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605837 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.606422 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data" (OuterVolumeSpecName: "config-data") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.606716 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.606739 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.607773 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.609491 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.609560 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd" (OuterVolumeSpecName: "kube-api-access-dr6pd") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "kube-api-access-dr6pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.626266 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.626648 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.627424 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.640170 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.708884 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709010 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dr6pd\" (UniqueName: \"kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709082 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709159 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709224 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709283 4824 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709342 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.723431 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.812173 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:55 crc kubenswrapper[4824]: I0121 11:59:55.147144 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"58d30951-ee56-4217-a614-a04951369c8b","Type":"ContainerDied","Data":"5bc5adacec79286d2293a6c562788392bcdca8f811b129442be3e2105c5568bf"} Jan 21 11:59:55 crc kubenswrapper[4824]: I0121 11:59:55.147356 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bc5adacec79286d2293a6c562788392bcdca8f811b129442be3e2105c5568bf" Jan 21 11:59:55 crc kubenswrapper[4824]: I0121 11:59:55.147204 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.145120 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj"] Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146267 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146282 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146301 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146307 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146322 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="extract-content" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146328 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="extract-content" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146344 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="extract-utilities" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146350 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="extract-utilities" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146358 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d30951-ee56-4217-a614-a04951369c8b" containerName="tempest-tests-tempest-tests-runner" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146364 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d30951-ee56-4217-a614-a04951369c8b" containerName="tempest-tests-tempest-tests-runner" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146386 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="extract-utilities" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146392 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="extract-utilities" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146399 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="extract-content" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146404 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="extract-content" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146600 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146629 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d30951-ee56-4217-a614-a04951369c8b" containerName="tempest-tests-tempest-tests-runner" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146641 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.147415 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.149468 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.149695 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.155569 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj"] Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.214313 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.214408 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hcrn\" (UniqueName: \"kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.214457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.315177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.315246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hcrn\" (UniqueName: \"kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.315277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.316381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.324853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.330753 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hcrn\" (UniqueName: \"kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.466176 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.867186 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj"] Jan 21 12:00:01 crc kubenswrapper[4824]: I0121 12:00:01.189853 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" event={"ID":"51822659-c0a9-4335-ad4d-8d1bff0a72fd","Type":"ContainerStarted","Data":"03ed238d87b1dad4b9d3e888042bef192d20cdc482bbe612d269cd50129896fd"} Jan 21 12:00:01 crc kubenswrapper[4824]: I0121 12:00:01.190186 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" event={"ID":"51822659-c0a9-4335-ad4d-8d1bff0a72fd","Type":"ContainerStarted","Data":"dd3f55cae47f30206c87faa984421a17438917a85679db9f01459ec6d110d380"} Jan 21 12:00:01 crc kubenswrapper[4824]: I0121 12:00:01.205584 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" podStartSLOduration=1.205569927 podStartE2EDuration="1.205569927s" podCreationTimestamp="2026-01-21 12:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 12:00:01.202116295 +0000 UTC m=+2943.495145588" watchObservedRunningTime="2026-01-21 12:00:01.205569927 +0000 UTC m=+2943.498599219" Jan 21 12:00:02 crc kubenswrapper[4824]: I0121 12:00:02.200438 4824 generic.go:334] "Generic (PLEG): container finished" podID="51822659-c0a9-4335-ad4d-8d1bff0a72fd" containerID="03ed238d87b1dad4b9d3e888042bef192d20cdc482bbe612d269cd50129896fd" exitCode=0 Jan 21 12:00:02 crc kubenswrapper[4824]: I0121 12:00:02.200489 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" event={"ID":"51822659-c0a9-4335-ad4d-8d1bff0a72fd","Type":"ContainerDied","Data":"03ed238d87b1dad4b9d3e888042bef192d20cdc482bbe612d269cd50129896fd"} Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.505564 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.573229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume\") pod \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.573297 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hcrn\" (UniqueName: \"kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn\") pod \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.573365 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume\") pod \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.574157 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume" (OuterVolumeSpecName: "config-volume") pod "51822659-c0a9-4335-ad4d-8d1bff0a72fd" (UID: "51822659-c0a9-4335-ad4d-8d1bff0a72fd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.579463 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "51822659-c0a9-4335-ad4d-8d1bff0a72fd" (UID: "51822659-c0a9-4335-ad4d-8d1bff0a72fd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.579677 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn" (OuterVolumeSpecName: "kube-api-access-7hcrn") pod "51822659-c0a9-4335-ad4d-8d1bff0a72fd" (UID: "51822659-c0a9-4335-ad4d-8d1bff0a72fd"). InnerVolumeSpecName "kube-api-access-7hcrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.676250 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.676285 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hcrn\" (UniqueName: \"kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn\") on node \"crc\" DevicePath \"\"" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.676298 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 12:00:04 crc kubenswrapper[4824]: I0121 12:00:04.218525 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" event={"ID":"51822659-c0a9-4335-ad4d-8d1bff0a72fd","Type":"ContainerDied","Data":"dd3f55cae47f30206c87faa984421a17438917a85679db9f01459ec6d110d380"} Jan 21 12:00:04 crc kubenswrapper[4824]: I0121 12:00:04.218575 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd3f55cae47f30206c87faa984421a17438917a85679db9f01459ec6d110d380" Jan 21 12:00:04 crc kubenswrapper[4824]: I0121 12:00:04.218622 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:04 crc kubenswrapper[4824]: I0121 12:00:04.278034 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5"] Jan 21 12:00:04 crc kubenswrapper[4824]: I0121 12:00:04.285979 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5"] Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.050011 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:00:05 crc kubenswrapper[4824]: E0121 12:00:05.051272 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.414333 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 21 12:00:05 crc kubenswrapper[4824]: E0121 12:00:05.414828 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51822659-c0a9-4335-ad4d-8d1bff0a72fd" containerName="collect-profiles" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.414847 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="51822659-c0a9-4335-ad4d-8d1bff0a72fd" containerName="collect-profiles" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.415118 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="51822659-c0a9-4335-ad4d-8d1bff0a72fd" containerName="collect-profiles" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.415795 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.418312 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qs54z" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.419497 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.611405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjl5c\" (UniqueName: \"kubernetes.io/projected/333160e2-3c05-401f-af89-abae8c0cb9d9-kube-api-access-cjl5c\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.611731 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.713106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjl5c\" (UniqueName: \"kubernetes.io/projected/333160e2-3c05-401f-af89-abae8c0cb9d9-kube-api-access-cjl5c\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.713175 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.713650 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.731629 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjl5c\" (UniqueName: \"kubernetes.io/projected/333160e2-3c05-401f-af89-abae8c0cb9d9-kube-api-access-cjl5c\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.734915 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:06 crc kubenswrapper[4824]: I0121 12:00:06.030795 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:06 crc kubenswrapper[4824]: I0121 12:00:06.061787 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="123c89e0-3113-4b28-9129-3848a7288052" path="/var/lib/kubelet/pods/123c89e0-3113-4b28-9129-3848a7288052/volumes" Jan 21 12:00:06 crc kubenswrapper[4824]: I0121 12:00:06.440644 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 21 12:00:07 crc kubenswrapper[4824]: I0121 12:00:07.249719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"333160e2-3c05-401f-af89-abae8c0cb9d9","Type":"ContainerStarted","Data":"7b61a3f10a73097a6b45f89bcec677347f8c66f5dfa2b3f35db827e53eab7185"} Jan 21 12:00:08 crc kubenswrapper[4824]: I0121 12:00:08.258867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"333160e2-3c05-401f-af89-abae8c0cb9d9","Type":"ContainerStarted","Data":"8032ae7e95d06d6b6e62f69d1cbf031c04ad5d6ec99a69cbbb9694bd7abedba3"} Jan 21 12:00:08 crc kubenswrapper[4824]: I0121 12:00:08.275032 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.325221253 podStartE2EDuration="3.275009063s" podCreationTimestamp="2026-01-21 12:00:05 +0000 UTC" firstStartedPulling="2026-01-21 12:00:06.443953385 +0000 UTC m=+2948.736982678" lastFinishedPulling="2026-01-21 12:00:07.393741196 +0000 UTC m=+2949.686770488" observedRunningTime="2026-01-21 12:00:08.269543549 +0000 UTC m=+2950.562572841" watchObservedRunningTime="2026-01-21 12:00:08.275009063 +0000 UTC m=+2950.568038355" Jan 21 12:00:17 crc kubenswrapper[4824]: I0121 12:00:17.050843 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:00:17 crc kubenswrapper[4824]: E0121 12:00:17.051611 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.038429 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-87fjw/must-gather-z5tfn"] Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.040583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.043913 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-87fjw"/"kube-root-ca.crt" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.049297 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-87fjw"/"openshift-service-ca.crt" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.066751 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-87fjw/must-gather-z5tfn"] Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.100079 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.100310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjtv2\" (UniqueName: \"kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.206544 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjtv2\" (UniqueName: \"kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.209709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.210220 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.260046 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjtv2\" (UniqueName: \"kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.357679 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: W0121 12:00:26.802668 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2245f836_eda4_4bf5_aa8a_5b18de2927b2.slice/crio-ba666897ba27d94dafdd68ff02864be8c54ee5955437ed8916a10c4f73931701 WatchSource:0}: Error finding container ba666897ba27d94dafdd68ff02864be8c54ee5955437ed8916a10c4f73931701: Status 404 returned error can't find the container with id ba666897ba27d94dafdd68ff02864be8c54ee5955437ed8916a10c4f73931701 Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.802952 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-87fjw/must-gather-z5tfn"] Jan 21 12:00:27 crc kubenswrapper[4824]: I0121 12:00:27.428937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/must-gather-z5tfn" event={"ID":"2245f836-eda4-4bf5-aa8a-5b18de2927b2","Type":"ContainerStarted","Data":"ba666897ba27d94dafdd68ff02864be8c54ee5955437ed8916a10c4f73931701"} Jan 21 12:00:31 crc kubenswrapper[4824]: I0121 12:00:31.049716 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:00:31 crc kubenswrapper[4824]: E0121 12:00:31.050542 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:00:33 crc kubenswrapper[4824]: I0121 12:00:33.495695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/must-gather-z5tfn" event={"ID":"2245f836-eda4-4bf5-aa8a-5b18de2927b2","Type":"ContainerStarted","Data":"a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320"} Jan 21 12:00:33 crc kubenswrapper[4824]: I0121 12:00:33.496481 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/must-gather-z5tfn" event={"ID":"2245f836-eda4-4bf5-aa8a-5b18de2927b2","Type":"ContainerStarted","Data":"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe"} Jan 21 12:00:33 crc kubenswrapper[4824]: I0121 12:00:33.518817 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-87fjw/must-gather-z5tfn" podStartSLOduration=1.94938882 podStartE2EDuration="7.518416432s" podCreationTimestamp="2026-01-21 12:00:26 +0000 UTC" firstStartedPulling="2026-01-21 12:00:26.805298386 +0000 UTC m=+2969.098327678" lastFinishedPulling="2026-01-21 12:00:32.374325999 +0000 UTC m=+2974.667355290" observedRunningTime="2026-01-21 12:00:33.512850889 +0000 UTC m=+2975.805880180" watchObservedRunningTime="2026-01-21 12:00:33.518416432 +0000 UTC m=+2975.811445724" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.635024 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-87fjw/crc-debug-cs7ct"] Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.637551 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.639493 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-87fjw"/"default-dockercfg-xkx7x" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.733806 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7mlv\" (UniqueName: \"kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.733898 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.837391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7mlv\" (UniqueName: \"kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.837505 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.838034 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.863832 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7mlv\" (UniqueName: \"kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.957666 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:36 crc kubenswrapper[4824]: I0121 12:00:36.522950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" event={"ID":"67011da1-4c52-4b94-8f0f-c6f8349af235","Type":"ContainerStarted","Data":"fc5413e9aaabe44b332afe86bee51d572012c0e71b94f4311343b145fb32e634"} Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.045173 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7f965fb4-8slr2_3f7f97b5-8abd-4351-8396-cf1b76c376a7/barbican-api-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.055399 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7f965fb4-8slr2_3f7f97b5-8abd-4351-8396-cf1b76c376a7/barbican-api/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.160545 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b7cfb8cd-ktpp5_8b84edbe-4485-4027-bfc4-cebbb1b10543/barbican-keystone-listener-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.166319 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b7cfb8cd-ktpp5_8b84edbe-4485-4027-bfc4-cebbb1b10543/barbican-keystone-listener/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.180838 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7cf777545f-f8drb_283e6982-7444-48c7-b7ac-ed4210329a08/barbican-worker-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.187213 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7cf777545f-f8drb_283e6982-7444-48c7-b7ac-ed4210329a08/barbican-worker/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.216468 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6_300a037e-ac25-4a9e-9339-6cc707969f7c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.256812 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/ceilometer-central-agent/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.278842 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/ceilometer-notification-agent/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.283392 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/sg-core/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.292430 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/proxy-httpd/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.307332 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ff9de537-ecd2-469d-b90e-5857bbf64c2d/cinder-api-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.336936 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ff9de537-ecd2-469d-b90e-5857bbf64c2d/cinder-api/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.364235 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e0f0267-58d2-4690-9187-8ddf03ef082e/cinder-scheduler/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.394653 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e0f0267-58d2-4690-9187-8ddf03ef082e/probe/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.416871 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz_b8c7ec7b-ee3f-46bc-95f8-4406713ba43a/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.436084 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n_8d46d1eb-4230-49c7-8376-698e7689ad2b/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.470767 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-26d8b_976a6446-e79c-42db-be45-43147767f1da/dnsmasq-dns/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.475201 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-26d8b_976a6446-e79c-42db-be45-43147767f1da/init/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.500849 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9_8dbaf72e-72a7-4073-a4c1-95cbdabaa64b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.515643 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_219725d0-9bef-4e42-8b78-cbc74a8ac9db/glance-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.532915 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_219725d0-9bef-4e42-8b78-cbc74a8ac9db/glance-httpd/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.543598 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7419ae4a-e8ea-4384-a9b8-3a3e02e27f25/glance-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.563609 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7419ae4a-e8ea-4384-a9b8-3a3e02e27f25/glance-httpd/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.005307 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-549445f5fc-fwxlz_8655cad5-2680-4cf6-b50a-0745adeeb469/heat-api/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.244169 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-68b8cf894c-ws5zd_a1aefb21-aac8-4f8f-a402-16006c1f336a/heat-cfnapi/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.364625 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5bc6fc99dd-nv45v_ca8530d7-bcfb-45fa-b0ff-a2702c081bda/heat-engine/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.385917 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2_b821ea8d-9327-4623-b5be-3da7d3872cab/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.414363 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-b74gh_60fedbf7-07e6-41be-ada3-2a06f28835e3/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.768293 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c5fc769c9-spmzr_7d943b5e-d05f-445a-888d-7ebe340468b9/keystone-api/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.780425 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_d42233a9-cfd9-486d-b314-f17d6a9d3b03/kube-state-metrics/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.815675 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn_fdde8991-c204-45a0-b344-dcdc41a9a275/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:43 crc kubenswrapper[4824]: I0121 12:00:43.050349 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:00:43 crc kubenswrapper[4824]: E0121 12:00:43.051138 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:00:47 crc kubenswrapper[4824]: I0121 12:00:47.633198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" event={"ID":"67011da1-4c52-4b94-8f0f-c6f8349af235","Type":"ContainerStarted","Data":"c3e2158290b1bee0a6e844d1cfc3b1ea1c8ca3e18fabfc983d0a39b05fe5af6d"} Jan 21 12:00:47 crc kubenswrapper[4824]: I0121 12:00:47.649586 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" podStartSLOduration=1.220151406 podStartE2EDuration="12.649570589s" podCreationTimestamp="2026-01-21 12:00:35 +0000 UTC" firstStartedPulling="2026-01-21 12:00:36.003670906 +0000 UTC m=+2978.296700198" lastFinishedPulling="2026-01-21 12:00:47.433090089 +0000 UTC m=+2989.726119381" observedRunningTime="2026-01-21 12:00:47.648243057 +0000 UTC m=+2989.941272359" watchObservedRunningTime="2026-01-21 12:00:47.649570589 +0000 UTC m=+2989.942599881" Jan 21 12:00:54 crc kubenswrapper[4824]: I0121 12:00:54.049368 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:00:54 crc kubenswrapper[4824]: E0121 12:00:54.050592 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.089342 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_dba0050b-8a73-4355-a1e0-7c9a03557ead/memcached/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.194635 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-659565f76f-jdg2q_b6f2ba03-746e-465c-bde2-c917129fb125/neutron-api/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.252817 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-659565f76f-jdg2q_b6f2ba03-746e-465c-bde2-c917129fb125/neutron-httpd/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.274019 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l_4f3c3a58-5d6b-4cc2-9707-d088b2fd463d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.482387 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55525bd3-ae2b-494d-9a33-dd7d00c576b2/nova-api-log/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.763104 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55525bd3-ae2b-494d-9a33-dd7d00c576b2/nova-api-api/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.860705 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8effdacf-8fec-40b4-8f61-8856ac0f232c/nova-cell0-conductor-conductor/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.934030 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c398d1e9-13e1-400c-a609-bceea49aea51/nova-cell1-conductor-conductor/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.983437 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8e42e9d6-485f-4b97-b049-316cd47af30e/nova-cell1-novncproxy-novncproxy/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.032610 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-7crw8_7355a8fc-8c18-4de1-8baf-f6294ea9e11c/nova-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.146502 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_654a41d0-b44a-4194-a514-a3c60126d0d9/nova-metadata-log/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.823499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_654a41d0-b44a-4194-a514-a3c60126d0d9/nova-metadata-metadata/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.922520 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_7860ffbb-b435-4736-b2d3-f8871e70dc7a/nova-scheduler-scheduler/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.943166 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cd8d62b-ab74-4a1b-95f5-7253ed6c0346/galera/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.954721 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cd8d62b-ab74-4a1b-95f5-7253ed6c0346/mysql-bootstrap/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.975725 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_218aca21-4c37-4862-b891-fc228baa5348/galera/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.987750 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_218aca21-4c37-4862-b891-fc228baa5348/mysql-bootstrap/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.996590 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1af94d08-6d1e-4a97-8f82-27a1755d17a1/openstackclient/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.010852 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-htwd6_511f2b6d-a08d-49f8-b393-ab222219d4a7/ovn-controller/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.020926 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vrpmw_8652f3f6-1ad7-49d9-9048-827be17a3dd0/openstack-network-exporter/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.034410 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovsdb-server/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.045322 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovs-vswitchd/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.052919 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovsdb-server-init/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.090327 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gf4zq_18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.100092 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b9c25b1-d0b9-430e-88cf-0aeaf6785461/ovn-northd/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.104664 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b9c25b1-d0b9-430e-88cf-0aeaf6785461/openstack-network-exporter/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.118863 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5c0b5876-da04-49e6-8946-4fc6060a64ec/ovsdbserver-nb/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.127455 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5c0b5876-da04-49e6-8946-4fc6060a64ec/openstack-network-exporter/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.139837 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef/ovsdbserver-sb/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.148098 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef/openstack-network-exporter/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.216846 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8654bbf998-p8crm_c23477ba-5a41-44da-a8b7-6606595b9726/placement-log/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.271722 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8654bbf998-p8crm_c23477ba-5a41-44da-a8b7-6606595b9726/placement-api/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.295540 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d8e256a-aa1b-4ccb-8e8c-c86be1e31762/rabbitmq/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.299177 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d8e256a-aa1b-4ccb-8e8c-c86be1e31762/setup-container/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.324599 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4a4d725-1a49-4292-af74-6ab09ab15cdb/rabbitmq/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.328714 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4a4d725-1a49-4292-af74-6ab09ab15cdb/setup-container/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.341701 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w_890acdab-9f31-4afc-80e4-c1df308ac5a6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.350996 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-2flzn_34c68ecf-cf29-44cb-9880-a3d45f625454/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.361391 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf_e3664d03-d53d-4a49-8cab-c4266f2d1426/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.370449 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pn4sw_ad533443-01b7-4860-9920-93ed67f6b52f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.380423 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-wbrwb_4529344f-3dfd-4858-83f1-4abffe47f2ad/ssh-known-hosts-edpm-deployment/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.463486 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-79f4c948b7-4hz4p_01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d/proxy-httpd/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.478080 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-79f4c948b7-4hz4p_01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d/proxy-server/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.486124 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-nc6vf_29473359-3517-4b86-bd5e-80e25706ff27/swift-ring-rebalance/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.509216 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-server/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.525284 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-replicator/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.529593 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-auditor/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.536101 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-reaper/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.545939 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-server/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.562450 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-replicator/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.566104 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-auditor/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.574170 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-updater/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.582348 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-server/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.596440 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-replicator/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.605243 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-auditor/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.613219 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-updater/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.623134 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-expirer/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.627057 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/rsync/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.633312 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/swift-recon-cron/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.693596 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6_49b2a136-e5c3-4741-80c6-4c545d4a82a0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.720571 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_58d30951-ee56-4217-a614-a04951369c8b/tempest-tests-tempest-tests-runner/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.729009 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_333160e2-3c05-401f-af89-abae8c0cb9d9/test-operator-logs-container/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.743933 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz_d672f153-6ef2-432d-829a-db63c8daab69/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:59 crc kubenswrapper[4824]: I0121 12:00:59.602479 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:00:59 crc kubenswrapper[4824]: I0121 12:00:59.608679 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:00:59 crc kubenswrapper[4824]: I0121 12:00:59.626302 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.149302 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29483281-bz8hk"] Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.158318 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29483281-bz8hk"] Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.158529 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.179212 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.179492 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.179542 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.179615 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5rns\" (UniqueName: \"kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.281546 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.281606 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.281671 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5rns\" (UniqueName: \"kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.281760 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.290623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.306351 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.312261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5rns\" (UniqueName: \"kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.319797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.487186 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.897259 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.907936 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.917684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.924526 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.932260 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.937781 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.951261 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.958298 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.969316 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.989744 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.997352 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.047140 4824 scope.go:117] "RemoveContainer" containerID="1c72418277c3fff2993fb8936f438b1d2049c9f13bdefd24ee7c2900f6904eca" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.066100 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29483281-bz8hk"] Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.369278 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.377051 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.744410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29483281-bz8hk" event={"ID":"c92d6f21-1fe2-41a2-a516-a0d0740ab36f","Type":"ContainerStarted","Data":"0f8293175001be831fd07c8e6b87e9d57d97179a89c7a7a13cc4a25f076d8348"} Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.744462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29483281-bz8hk" event={"ID":"c92d6f21-1fe2-41a2-a516-a0d0740ab36f","Type":"ContainerStarted","Data":"afdbd9e5b38ea1c61d377b1aa74001c97326aa386ba38e7c7774fbbd18823a66"} Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.759240 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29483281-bz8hk" podStartSLOduration=1.7592209479999998 podStartE2EDuration="1.759220948s" podCreationTimestamp="2026-01-21 12:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 12:01:01.75765157 +0000 UTC m=+3004.050680862" watchObservedRunningTime="2026-01-21 12:01:01.759220948 +0000 UTC m=+3004.052250240" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.888699 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.934450 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.943357 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.951907 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.964573 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.984405 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.056416 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.116735 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.124746 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.371183 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.381671 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.453548 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.462972 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.493579 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.534605 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.606530 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.616805 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.628626 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.784206 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.762286 4824 generic.go:334] "Generic (PLEG): container finished" podID="c92d6f21-1fe2-41a2-a516-a0d0740ab36f" containerID="0f8293175001be831fd07c8e6b87e9d57d97179a89c7a7a13cc4a25f076d8348" exitCode=0 Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.762336 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29483281-bz8hk" event={"ID":"c92d6f21-1fe2-41a2-a516-a0d0740ab36f","Type":"ContainerDied","Data":"0f8293175001be831fd07c8e6b87e9d57d97179a89c7a7a13cc4a25f076d8348"} Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.877025 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.884170 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.932332 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.953621 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.973526 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.993578 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:01:04 crc kubenswrapper[4824]: I0121 12:01:04.049503 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:01:04 crc kubenswrapper[4824]: I0121 12:01:04.063146 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:01:04 crc kubenswrapper[4824]: I0121 12:01:04.069531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.048432 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.113054 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5rns\" (UniqueName: \"kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns\") pod \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.113374 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle\") pod \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.113425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data\") pod \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.113497 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys\") pod \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.120090 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns" (OuterVolumeSpecName: "kube-api-access-l5rns") pod "c92d6f21-1fe2-41a2-a516-a0d0740ab36f" (UID: "c92d6f21-1fe2-41a2-a516-a0d0740ab36f"). InnerVolumeSpecName "kube-api-access-l5rns". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.125065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c92d6f21-1fe2-41a2-a516-a0d0740ab36f" (UID: "c92d6f21-1fe2-41a2-a516-a0d0740ab36f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.164314 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c92d6f21-1fe2-41a2-a516-a0d0740ab36f" (UID: "c92d6f21-1fe2-41a2-a516-a0d0740ab36f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.178775 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data" (OuterVolumeSpecName: "config-data") pod "c92d6f21-1fe2-41a2-a516-a0d0740ab36f" (UID: "c92d6f21-1fe2-41a2-a516-a0d0740ab36f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.225215 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.225241 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.225251 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.225263 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5rns\" (UniqueName: \"kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.779594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29483281-bz8hk" event={"ID":"c92d6f21-1fe2-41a2-a516-a0d0740ab36f","Type":"ContainerDied","Data":"afdbd9e5b38ea1c61d377b1aa74001c97326aa386ba38e7c7774fbbd18823a66"} Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.779654 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afdbd9e5b38ea1c61d377b1aa74001c97326aa386ba38e7c7774fbbd18823a66" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.779678 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:05 crc kubenswrapper[4824]: E0121 12:01:05.976669 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc92d6f21_1fe2_41a2_a516_a0d0740ab36f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc92d6f21_1fe2_41a2_a516_a0d0740ab36f.slice/crio-afdbd9e5b38ea1c61d377b1aa74001c97326aa386ba38e7c7774fbbd18823a66\": RecentStats: unable to find data in memory cache]" Jan 21 12:01:06 crc kubenswrapper[4824]: I0121 12:01:06.050489 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:01:06 crc kubenswrapper[4824]: E0121 12:01:06.051366 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:01:09 crc kubenswrapper[4824]: I0121 12:01:09.116639 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-r65dm_c117483a-fe5c-423b-995e-9c2e7f59566a/control-plane-machine-set-operator/0.log" Jan 21 12:01:09 crc kubenswrapper[4824]: I0121 12:01:09.129430 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/kube-rbac-proxy/0.log" Jan 21 12:01:09 crc kubenswrapper[4824]: I0121 12:01:09.137531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/machine-api-operator/0.log" Jan 21 12:01:20 crc kubenswrapper[4824]: I0121 12:01:20.061560 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:01:20 crc kubenswrapper[4824]: E0121 12:01:20.062723 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:01:22 crc kubenswrapper[4824]: I0121 12:01:22.929371 4824 generic.go:334] "Generic (PLEG): container finished" podID="67011da1-4c52-4b94-8f0f-c6f8349af235" containerID="c3e2158290b1bee0a6e844d1cfc3b1ea1c8ca3e18fabfc983d0a39b05fe5af6d" exitCode=0 Jan 21 12:01:22 crc kubenswrapper[4824]: I0121 12:01:22.929455 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" event={"ID":"67011da1-4c52-4b94-8f0f-c6f8349af235","Type":"ContainerDied","Data":"c3e2158290b1bee0a6e844d1cfc3b1ea1c8ca3e18fabfc983d0a39b05fe5af6d"} Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.012313 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.037023 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7mlv\" (UniqueName: \"kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv\") pod \"67011da1-4c52-4b94-8f0f-c6f8349af235\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.037192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host\") pod \"67011da1-4c52-4b94-8f0f-c6f8349af235\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.037322 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host" (OuterVolumeSpecName: "host") pod "67011da1-4c52-4b94-8f0f-c6f8349af235" (UID: "67011da1-4c52-4b94-8f0f-c6f8349af235"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.038151 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.041427 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-cs7ct"] Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.043589 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv" (OuterVolumeSpecName: "kube-api-access-c7mlv") pod "67011da1-4c52-4b94-8f0f-c6f8349af235" (UID: "67011da1-4c52-4b94-8f0f-c6f8349af235"). InnerVolumeSpecName "kube-api-access-c7mlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.058744 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-cs7ct"] Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.140345 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7mlv\" (UniqueName: \"kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.950418 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc5413e9aaabe44b332afe86bee51d572012c0e71b94f4311343b145fb32e634" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.950486 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.181891 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-87fjw/crc-debug-jqpvk"] Jan 21 12:01:25 crc kubenswrapper[4824]: E0121 12:01:25.182599 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92d6f21-1fe2-41a2-a516-a0d0740ab36f" containerName="keystone-cron" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.182616 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92d6f21-1fe2-41a2-a516-a0d0740ab36f" containerName="keystone-cron" Jan 21 12:01:25 crc kubenswrapper[4824]: E0121 12:01:25.182635 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67011da1-4c52-4b94-8f0f-c6f8349af235" containerName="container-00" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.182641 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="67011da1-4c52-4b94-8f0f-c6f8349af235" containerName="container-00" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.182885 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="67011da1-4c52-4b94-8f0f-c6f8349af235" containerName="container-00" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.182903 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c92d6f21-1fe2-41a2-a516-a0d0740ab36f" containerName="keystone-cron" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.183553 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.185414 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-87fjw"/"default-dockercfg-xkx7x" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.258186 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnpcq\" (UniqueName: \"kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.258520 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.359478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnpcq\" (UniqueName: \"kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.359562 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.359683 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.377403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnpcq\" (UniqueName: \"kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.501710 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.963989 4824 generic.go:334] "Generic (PLEG): container finished" podID="87698512-b8b0-416a-b593-d9ca8bc2eeac" containerID="c6793d88c2985e5e3d28f1236121b047a23b50c741fdeb44a8e3cc182cf06683" exitCode=0 Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.964076 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" event={"ID":"87698512-b8b0-416a-b593-d9ca8bc2eeac","Type":"ContainerDied","Data":"c6793d88c2985e5e3d28f1236121b047a23b50c741fdeb44a8e3cc182cf06683"} Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.964369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" event={"ID":"87698512-b8b0-416a-b593-d9ca8bc2eeac","Type":"ContainerStarted","Data":"a448dccd21cb3b27e3293d09c067735c54887c539a7fd0b2a40524c5960f4e56"} Jan 21 12:01:26 crc kubenswrapper[4824]: I0121 12:01:26.058610 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67011da1-4c52-4b94-8f0f-c6f8349af235" path="/var/lib/kubelet/pods/67011da1-4c52-4b94-8f0f-c6f8349af235/volumes" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.057834 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.195090 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host\") pod \"87698512-b8b0-416a-b593-d9ca8bc2eeac\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.195191 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host" (OuterVolumeSpecName: "host") pod "87698512-b8b0-416a-b593-d9ca8bc2eeac" (UID: "87698512-b8b0-416a-b593-d9ca8bc2eeac"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.195364 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnpcq\" (UniqueName: \"kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq\") pod \"87698512-b8b0-416a-b593-d9ca8bc2eeac\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.196995 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.201557 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq" (OuterVolumeSpecName: "kube-api-access-dnpcq") pod "87698512-b8b0-416a-b593-d9ca8bc2eeac" (UID: "87698512-b8b0-416a-b593-d9ca8bc2eeac"). InnerVolumeSpecName "kube-api-access-dnpcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.298597 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnpcq\" (UniqueName: \"kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.988533 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" event={"ID":"87698512-b8b0-416a-b593-d9ca8bc2eeac","Type":"ContainerDied","Data":"a448dccd21cb3b27e3293d09c067735c54887c539a7fd0b2a40524c5960f4e56"} Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.989077 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a448dccd21cb3b27e3293d09c067735c54887c539a7fd0b2a40524c5960f4e56" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.988597 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:28 crc kubenswrapper[4824]: I0121 12:01:28.114412 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-jqpvk"] Jan 21 12:01:28 crc kubenswrapper[4824]: I0121 12:01:28.123021 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-jqpvk"] Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.272359 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-87fjw/crc-debug-7b2zl"] Jan 21 12:01:29 crc kubenswrapper[4824]: E0121 12:01:29.272857 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87698512-b8b0-416a-b593-d9ca8bc2eeac" containerName="container-00" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.272871 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="87698512-b8b0-416a-b593-d9ca8bc2eeac" containerName="container-00" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.273122 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="87698512-b8b0-416a-b593-d9ca8bc2eeac" containerName="container-00" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.273780 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.276067 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-87fjw"/"default-dockercfg-xkx7x" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.434433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.434516 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdwzx\" (UniqueName: \"kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.536377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdwzx\" (UniqueName: \"kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.536552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.536633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.556402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdwzx\" (UniqueName: \"kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.595941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: W0121 12:01:29.622993 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad457817_57db_426b_8191_f35e1d82d266.slice/crio-ae1325f602a0dbb91e07fcc9773e501ad40e003b6336ac64860e84a0cb84054f WatchSource:0}: Error finding container ae1325f602a0dbb91e07fcc9773e501ad40e003b6336ac64860e84a0cb84054f: Status 404 returned error can't find the container with id ae1325f602a0dbb91e07fcc9773e501ad40e003b6336ac64860e84a0cb84054f Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.010467 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad457817-57db-426b-8191-f35e1d82d266" containerID="6e4471e366c5d457dbdb31676ebe2e10113437d05155002808726047c10699c0" exitCode=0 Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.010575 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" event={"ID":"ad457817-57db-426b-8191-f35e1d82d266","Type":"ContainerDied","Data":"6e4471e366c5d457dbdb31676ebe2e10113437d05155002808726047c10699c0"} Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.010899 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" event={"ID":"ad457817-57db-426b-8191-f35e1d82d266","Type":"ContainerStarted","Data":"ae1325f602a0dbb91e07fcc9773e501ad40e003b6336ac64860e84a0cb84054f"} Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.060973 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87698512-b8b0-416a-b593-d9ca8bc2eeac" path="/var/lib/kubelet/pods/87698512-b8b0-416a-b593-d9ca8bc2eeac/volumes" Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.061771 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-7b2zl"] Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.061818 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-7b2zl"] Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.516222 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.528035 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.535795 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.112567 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.168067 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdwzx\" (UniqueName: \"kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx\") pod \"ad457817-57db-426b-8191-f35e1d82d266\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.168141 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host\") pod \"ad457817-57db-426b-8191-f35e1d82d266\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.168381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host" (OuterVolumeSpecName: "host") pod "ad457817-57db-426b-8191-f35e1d82d266" (UID: "ad457817-57db-426b-8191-f35e1d82d266"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.168682 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.174200 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx" (OuterVolumeSpecName: "kube-api-access-bdwzx") pod "ad457817-57db-426b-8191-f35e1d82d266" (UID: "ad457817-57db-426b-8191-f35e1d82d266"). InnerVolumeSpecName "kube-api-access-bdwzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.270583 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdwzx\" (UniqueName: \"kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:32 crc kubenswrapper[4824]: I0121 12:01:32.039587 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae1325f602a0dbb91e07fcc9773e501ad40e003b6336ac64860e84a0cb84054f" Jan 21 12:01:32 crc kubenswrapper[4824]: I0121 12:01:32.039667 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:32 crc kubenswrapper[4824]: I0121 12:01:32.073725 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad457817-57db-426b-8191-f35e1d82d266" path="/var/lib/kubelet/pods/ad457817-57db-426b-8191-f35e1d82d266/volumes" Jan 21 12:01:34 crc kubenswrapper[4824]: I0121 12:01:34.050548 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:01:34 crc kubenswrapper[4824]: E0121 12:01:34.051541 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.213499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-xjqs5_1575ab9d-016f-403a-990b-3b4c4d53f6f1/nmstate-console-plugin/0.log" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.238527 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qjpv9_a01ecff0-9051-4ddb-8c8e-a4f379b6c823/nmstate-handler/0.log" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.251800 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/nmstate-metrics/0.log" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.257456 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/kube-rbac-proxy/0.log" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.273024 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-zmc4k_df60ca53-0573-4a6a-9235-86e0a1c17f91/nmstate-operator/0.log" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.291562 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-h6kkh_edf54909-b53e-44de-90bc-355e23bbc767/nmstate-webhook/0.log" Jan 21 12:01:43 crc kubenswrapper[4824]: I0121 12:01:43.221356 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:01:43 crc kubenswrapper[4824]: I0121 12:01:43.226091 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:01:43 crc kubenswrapper[4824]: I0121 12:01:43.242505 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.400325 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.412242 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.417328 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.422836 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.428295 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.436054 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.439443 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.445618 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.454484 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.473539 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.479924 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.768475 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.775246 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.505194 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/extract/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.514143 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/util/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.521256 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/pull/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.529505 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/extract/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.537776 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/util/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.544123 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/pull/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.868402 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/registry-server/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.873301 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/extract-utilities/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.881060 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/extract-content/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.054237 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:01:48 crc kubenswrapper[4824]: E0121 12:01:48.054591 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.320057 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/registry-server/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.325123 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/extract-utilities/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.331470 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/extract-content/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.347829 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xcxs9_b4c29122-d9b2-4f85-8b29-c0e7d888667f/marketplace-operator/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.456228 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/registry-server/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.460487 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/extract-utilities/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.465951 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/extract-content/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.924605 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/registry-server/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.929259 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/extract-utilities/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.936054 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/extract-content/0.log" Jan 21 12:02:02 crc kubenswrapper[4824]: I0121 12:02:02.049885 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:02:02 crc kubenswrapper[4824]: E0121 12:02:02.050850 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:02:16 crc kubenswrapper[4824]: I0121 12:02:16.049239 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:02:16 crc kubenswrapper[4824]: E0121 12:02:16.050104 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:02:29 crc kubenswrapper[4824]: I0121 12:02:29.049402 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:02:29 crc kubenswrapper[4824]: E0121 12:02:29.050416 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:02:42 crc kubenswrapper[4824]: I0121 12:02:42.050030 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:02:42 crc kubenswrapper[4824]: E0121 12:02:42.050706 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:02:47 crc kubenswrapper[4824]: I0121 12:02:47.977836 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:02:47 crc kubenswrapper[4824]: I0121 12:02:47.984043 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:02:47 crc kubenswrapper[4824]: I0121 12:02:47.999206 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:02:48 crc kubenswrapper[4824]: I0121 12:02:48.072728 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:02:48 crc kubenswrapper[4824]: I0121 12:02:48.089884 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:02:48 crc kubenswrapper[4824]: I0121 12:02:48.098307 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.079146 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.142035 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.151939 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.164762 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.178102 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.195640 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.306470 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.384378 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.392119 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.532220 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.546064 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.551994 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.561569 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.570214 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.577929 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.584410 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.600248 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.606333 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.639345 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.647385 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.717967 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.733345 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.831224 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.841673 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.882737 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.941247 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.070277 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.079792 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.093470 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.177377 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.189323 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.273038 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.125814 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.142338 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.148943 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.399860 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.413512 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.468970 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.494743 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.515692 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.539930 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.595290 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.604565 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.617377 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.804864 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-r65dm_c117483a-fe5c-423b-995e-9c2e7f59566a/control-plane-machine-set-operator/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.815813 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/kube-rbac-proxy/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.823341 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/machine-api-operator/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.548436 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.578673 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.586872 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.593107 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.601511 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.613141 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.703207 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.790599 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.800750 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.801650 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-xjqs5_1575ab9d-016f-403a-990b-3b4c4d53f6f1/nmstate-console-plugin/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.824517 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qjpv9_a01ecff0-9051-4ddb-8c8e-a4f379b6c823/nmstate-handler/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.833467 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/nmstate-metrics/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.840689 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/kube-rbac-proxy/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.855586 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-zmc4k_df60ca53-0573-4a6a-9235-86e0a1c17f91/nmstate-operator/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.875637 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-h6kkh_edf54909-b53e-44de-90bc-355e23bbc767/nmstate-webhook/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.057451 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.068490 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.126023 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.137350 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.170507 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.208261 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.283860 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.296612 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.312344 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.470838 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.061771 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:02:54 crc kubenswrapper[4824]: E0121 12:02:54.062422 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.720502 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.734072 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.789840 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.812317 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.829579 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.850330 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.908385 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.919569 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.928319 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.325700 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/kube-multus-additional-cni-plugins/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.332998 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/egress-router-binary-copy/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.340184 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/cni-plugins/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.344861 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/bond-cni-plugin/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.349947 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/routeoverride-cni/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.355371 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/whereabouts-cni-bincopy/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.361608 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/whereabouts-cni/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.391859 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-27d2q_7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34/multus-admission-controller/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.395772 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-27d2q_7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34/kube-rbac-proxy/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.441794 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/2.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.534438 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/3.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.569925 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-9rmnw_8f9bd12b-22ee-44ed-a8b9-c100d043f691/network-metrics-daemon/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.575263 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-9rmnw_8f9bd12b-22ee-44ed-a8b9-c100d043f691/kube-rbac-proxy/0.log" Jan 21 12:03:08 crc kubenswrapper[4824]: I0121 12:03:08.056229 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:03:08 crc kubenswrapper[4824]: E0121 12:03:08.057375 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:03:22 crc kubenswrapper[4824]: I0121 12:03:22.049888 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:03:22 crc kubenswrapper[4824]: E0121 12:03:22.050849 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:03:37 crc kubenswrapper[4824]: I0121 12:03:37.048900 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:03:37 crc kubenswrapper[4824]: E0121 12:03:37.049698 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.179810 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:40 crc kubenswrapper[4824]: E0121 12:03:40.180672 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad457817-57db-426b-8191-f35e1d82d266" containerName="container-00" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.180685 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad457817-57db-426b-8191-f35e1d82d266" containerName="container-00" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.180884 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad457817-57db-426b-8191-f35e1d82d266" containerName="container-00" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.186533 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.195091 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.368500 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.368584 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.369070 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxf8c\" (UniqueName: \"kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.471656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxf8c\" (UniqueName: \"kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.471812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.471842 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.472371 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.472398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.492852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxf8c\" (UniqueName: \"kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.510933 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:41 crc kubenswrapper[4824]: I0121 12:03:41.066125 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:41 crc kubenswrapper[4824]: I0121 12:03:41.089643 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerStarted","Data":"6c39dd25e15891de0d35d968c22d1ed84039fb4585c5ab11eeb1d6baac169fee"} Jan 21 12:03:42 crc kubenswrapper[4824]: I0121 12:03:42.100254 4824 generic.go:334] "Generic (PLEG): container finished" podID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerID="717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff" exitCode=0 Jan 21 12:03:42 crc kubenswrapper[4824]: I0121 12:03:42.100489 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerDied","Data":"717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff"} Jan 21 12:03:42 crc kubenswrapper[4824]: I0121 12:03:42.103252 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 12:03:44 crc kubenswrapper[4824]: I0121 12:03:44.117362 4824 generic.go:334] "Generic (PLEG): container finished" podID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerID="3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0" exitCode=0 Jan 21 12:03:44 crc kubenswrapper[4824]: I0121 12:03:44.117542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerDied","Data":"3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0"} Jan 21 12:03:45 crc kubenswrapper[4824]: I0121 12:03:45.130612 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerStarted","Data":"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84"} Jan 21 12:03:45 crc kubenswrapper[4824]: I0121 12:03:45.171393 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rv2rv" podStartSLOduration=2.616691652 podStartE2EDuration="5.171364388s" podCreationTimestamp="2026-01-21 12:03:40 +0000 UTC" firstStartedPulling="2026-01-21 12:03:42.10291993 +0000 UTC m=+3164.395949223" lastFinishedPulling="2026-01-21 12:03:44.657592667 +0000 UTC m=+3166.950621959" observedRunningTime="2026-01-21 12:03:45.169929714 +0000 UTC m=+3167.462959006" watchObservedRunningTime="2026-01-21 12:03:45.171364388 +0000 UTC m=+3167.464393671" Jan 21 12:03:50 crc kubenswrapper[4824]: I0121 12:03:50.048927 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:03:50 crc kubenswrapper[4824]: E0121 12:03:50.050572 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:03:50 crc kubenswrapper[4824]: I0121 12:03:50.511303 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:50 crc kubenswrapper[4824]: I0121 12:03:50.511668 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:50 crc kubenswrapper[4824]: I0121 12:03:50.549330 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:51 crc kubenswrapper[4824]: I0121 12:03:51.233744 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:51 crc kubenswrapper[4824]: I0121 12:03:51.280156 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.214717 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rv2rv" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="registry-server" containerID="cri-o://15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84" gracePeriod=2 Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.644044 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.821047 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxf8c\" (UniqueName: \"kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c\") pod \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.821142 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities\") pod \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.821176 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content\") pod \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.821810 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities" (OuterVolumeSpecName: "utilities") pod "f22ac107-20fa-4273-9c3a-0d0d41b26c8f" (UID: "f22ac107-20fa-4273-9c3a-0d0d41b26c8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.822511 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.827867 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c" (OuterVolumeSpecName: "kube-api-access-nxf8c") pod "f22ac107-20fa-4273-9c3a-0d0d41b26c8f" (UID: "f22ac107-20fa-4273-9c3a-0d0d41b26c8f"). InnerVolumeSpecName "kube-api-access-nxf8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.857947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f22ac107-20fa-4273-9c3a-0d0d41b26c8f" (UID: "f22ac107-20fa-4273-9c3a-0d0d41b26c8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.925640 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxf8c\" (UniqueName: \"kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c\") on node \"crc\" DevicePath \"\"" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.925756 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.224136 4824 generic.go:334] "Generic (PLEG): container finished" podID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerID="15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84" exitCode=0 Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.224182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerDied","Data":"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84"} Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.224198 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.224222 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerDied","Data":"6c39dd25e15891de0d35d968c22d1ed84039fb4585c5ab11eeb1d6baac169fee"} Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.224248 4824 scope.go:117] "RemoveContainer" containerID="15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.251761 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.252436 4824 scope.go:117] "RemoveContainer" containerID="3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.260926 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.269995 4824 scope.go:117] "RemoveContainer" containerID="717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.307137 4824 scope.go:117] "RemoveContainer" containerID="15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84" Jan 21 12:03:54 crc kubenswrapper[4824]: E0121 12:03:54.307421 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84\": container with ID starting with 15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84 not found: ID does not exist" containerID="15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.307458 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84"} err="failed to get container status \"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84\": rpc error: code = NotFound desc = could not find container \"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84\": container with ID starting with 15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84 not found: ID does not exist" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.307484 4824 scope.go:117] "RemoveContainer" containerID="3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0" Jan 21 12:03:54 crc kubenswrapper[4824]: E0121 12:03:54.307742 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0\": container with ID starting with 3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0 not found: ID does not exist" containerID="3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.307764 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0"} err="failed to get container status \"3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0\": rpc error: code = NotFound desc = could not find container \"3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0\": container with ID starting with 3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0 not found: ID does not exist" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.307779 4824 scope.go:117] "RemoveContainer" containerID="717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff" Jan 21 12:03:54 crc kubenswrapper[4824]: E0121 12:03:54.308030 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff\": container with ID starting with 717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff not found: ID does not exist" containerID="717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.308060 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff"} err="failed to get container status \"717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff\": rpc error: code = NotFound desc = could not find container \"717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff\": container with ID starting with 717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff not found: ID does not exist" Jan 21 12:03:56 crc kubenswrapper[4824]: I0121 12:03:56.060226 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" path="/var/lib/kubelet/pods/f22ac107-20fa-4273-9c3a-0d0d41b26c8f/volumes" Jan 21 12:04:05 crc kubenswrapper[4824]: I0121 12:04:05.050216 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:04:05 crc kubenswrapper[4824]: E0121 12:04:05.051387 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:04:17 crc kubenswrapper[4824]: I0121 12:04:17.051047 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:04:17 crc kubenswrapper[4824]: I0121 12:04:17.424743 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d"} Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.134761 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:04:43 crc kubenswrapper[4824]: E0121 12:04:43.136315 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="registry-server" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.136336 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="registry-server" Jan 21 12:04:43 crc kubenswrapper[4824]: E0121 12:04:43.136379 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="extract-utilities" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.136386 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="extract-utilities" Jan 21 12:04:43 crc kubenswrapper[4824]: E0121 12:04:43.136399 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="extract-content" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.136405 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="extract-content" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.136724 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="registry-server" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.138977 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.151590 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.229541 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.229744 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bkcw\" (UniqueName: \"kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.229800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.331708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.331812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bkcw\" (UniqueName: \"kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.331835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.332299 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.332297 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.356848 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bkcw\" (UniqueName: \"kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.461514 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.917690 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:04:44 crc kubenswrapper[4824]: I0121 12:04:44.688036 4824 generic.go:334] "Generic (PLEG): container finished" podID="92367417-426c-47d7-ba64-b54f204fa7eb" containerID="bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a" exitCode=0 Jan 21 12:04:44 crc kubenswrapper[4824]: I0121 12:04:44.688156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerDied","Data":"bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a"} Jan 21 12:04:44 crc kubenswrapper[4824]: I0121 12:04:44.688426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerStarted","Data":"a7db3ead5f33a027665a2a78595a34e4bf12954ed323508aab571a54c6300caf"} Jan 21 12:04:45 crc kubenswrapper[4824]: I0121 12:04:45.698796 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerStarted","Data":"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a"} Jan 21 12:04:48 crc kubenswrapper[4824]: I0121 12:04:48.731127 4824 generic.go:334] "Generic (PLEG): container finished" podID="92367417-426c-47d7-ba64-b54f204fa7eb" containerID="c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a" exitCode=0 Jan 21 12:04:48 crc kubenswrapper[4824]: I0121 12:04:48.731321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerDied","Data":"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a"} Jan 21 12:04:49 crc kubenswrapper[4824]: I0121 12:04:49.744256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerStarted","Data":"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7"} Jan 21 12:04:49 crc kubenswrapper[4824]: I0121 12:04:49.770308 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hwgw5" podStartSLOduration=2.078753089 podStartE2EDuration="6.770284038s" podCreationTimestamp="2026-01-21 12:04:43 +0000 UTC" firstStartedPulling="2026-01-21 12:04:44.690179002 +0000 UTC m=+3226.983208294" lastFinishedPulling="2026-01-21 12:04:49.38170995 +0000 UTC m=+3231.674739243" observedRunningTime="2026-01-21 12:04:49.760778578 +0000 UTC m=+3232.053807869" watchObservedRunningTime="2026-01-21 12:04:49.770284038 +0000 UTC m=+3232.063313330" Jan 21 12:04:53 crc kubenswrapper[4824]: I0121 12:04:53.462063 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:53 crc kubenswrapper[4824]: I0121 12:04:53.462782 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:54 crc kubenswrapper[4824]: I0121 12:04:54.496659 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hwgw5" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="registry-server" probeResult="failure" output=< Jan 21 12:04:54 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Jan 21 12:04:54 crc kubenswrapper[4824]: > Jan 21 12:05:03 crc kubenswrapper[4824]: I0121 12:05:03.504561 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:05:03 crc kubenswrapper[4824]: I0121 12:05:03.549614 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:05:03 crc kubenswrapper[4824]: I0121 12:05:03.746281 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:05:04 crc kubenswrapper[4824]: I0121 12:05:04.870262 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hwgw5" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="registry-server" containerID="cri-o://09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7" gracePeriod=2 Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.301393 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.495252 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bkcw\" (UniqueName: \"kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw\") pod \"92367417-426c-47d7-ba64-b54f204fa7eb\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.495355 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content\") pod \"92367417-426c-47d7-ba64-b54f204fa7eb\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.495388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities\") pod \"92367417-426c-47d7-ba64-b54f204fa7eb\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.496262 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities" (OuterVolumeSpecName: "utilities") pod "92367417-426c-47d7-ba64-b54f204fa7eb" (UID: "92367417-426c-47d7-ba64-b54f204fa7eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.504037 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw" (OuterVolumeSpecName: "kube-api-access-8bkcw") pod "92367417-426c-47d7-ba64-b54f204fa7eb" (UID: "92367417-426c-47d7-ba64-b54f204fa7eb"). InnerVolumeSpecName "kube-api-access-8bkcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.594475 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92367417-426c-47d7-ba64-b54f204fa7eb" (UID: "92367417-426c-47d7-ba64-b54f204fa7eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.598593 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bkcw\" (UniqueName: \"kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw\") on node \"crc\" DevicePath \"\"" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.598635 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.598648 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.881414 4824 generic.go:334] "Generic (PLEG): container finished" podID="92367417-426c-47d7-ba64-b54f204fa7eb" containerID="09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7" exitCode=0 Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.881456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerDied","Data":"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7"} Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.881523 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.881550 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerDied","Data":"a7db3ead5f33a027665a2a78595a34e4bf12954ed323508aab571a54c6300caf"} Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.881600 4824 scope.go:117] "RemoveContainer" containerID="09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.934051 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.936166 4824 scope.go:117] "RemoveContainer" containerID="c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.938098 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.956148 4824 scope.go:117] "RemoveContainer" containerID="bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.987473 4824 scope.go:117] "RemoveContainer" containerID="09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7" Jan 21 12:05:05 crc kubenswrapper[4824]: E0121 12:05:05.987784 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7\": container with ID starting with 09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7 not found: ID does not exist" containerID="09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.987814 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7"} err="failed to get container status \"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7\": rpc error: code = NotFound desc = could not find container \"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7\": container with ID starting with 09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7 not found: ID does not exist" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.987837 4824 scope.go:117] "RemoveContainer" containerID="c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a" Jan 21 12:05:05 crc kubenswrapper[4824]: E0121 12:05:05.988145 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a\": container with ID starting with c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a not found: ID does not exist" containerID="c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.988209 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a"} err="failed to get container status \"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a\": rpc error: code = NotFound desc = could not find container \"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a\": container with ID starting with c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a not found: ID does not exist" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.988256 4824 scope.go:117] "RemoveContainer" containerID="bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a" Jan 21 12:05:05 crc kubenswrapper[4824]: E0121 12:05:05.988522 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a\": container with ID starting with bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a not found: ID does not exist" containerID="bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.988557 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a"} err="failed to get container status \"bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a\": rpc error: code = NotFound desc = could not find container \"bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a\": container with ID starting with bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a not found: ID does not exist" Jan 21 12:05:06 crc kubenswrapper[4824]: I0121 12:05:06.061058 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" path="/var/lib/kubelet/pods/92367417-426c-47d7-ba64-b54f204fa7eb/volumes" Jan 21 12:06:46 crc kubenswrapper[4824]: I0121 12:06:46.065345 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:06:46 crc kubenswrapper[4824]: I0121 12:06:46.067990 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:07:01 crc kubenswrapper[4824]: I0121 12:07:01.324702 4824 scope.go:117] "RemoveContainer" containerID="c3e2158290b1bee0a6e844d1cfc3b1ea1c8ca3e18fabfc983d0a39b05fe5af6d" Jan 21 12:07:16 crc kubenswrapper[4824]: I0121 12:07:16.064751 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:07:16 crc kubenswrapper[4824]: I0121 12:07:16.065477 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.065263 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.065835 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.065904 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.066806 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.066875 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d" gracePeriod=600 Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.357577 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d" exitCode=0 Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.357658 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d"} Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.357891 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:07:47 crc kubenswrapper[4824]: I0121 12:07:47.368122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4"} Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.145040 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:07:52 crc kubenswrapper[4824]: E0121 12:07:52.146500 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="extract-content" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.146521 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="extract-content" Jan 21 12:07:52 crc kubenswrapper[4824]: E0121 12:07:52.146585 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="extract-utilities" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.146593 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="extract-utilities" Jan 21 12:07:52 crc kubenswrapper[4824]: E0121 12:07:52.146614 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="registry-server" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.146624 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="registry-server" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.147598 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="registry-server" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.150057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.155905 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.199265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.199935 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.200204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf6dz\" (UniqueName: \"kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.301591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.301937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf6dz\" (UniqueName: \"kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.302315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.302400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.302637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.323336 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf6dz\" (UniqueName: \"kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.470811 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.895678 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:07:53 crc kubenswrapper[4824]: I0121 12:07:53.426011 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerID="5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59" exitCode=0 Jan 21 12:07:53 crc kubenswrapper[4824]: I0121 12:07:53.426206 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerDied","Data":"5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59"} Jan 21 12:07:53 crc kubenswrapper[4824]: I0121 12:07:53.427298 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerStarted","Data":"736f0aba8b28cb81d6e6ab52fa2cf14cf8d30b070cc2c19d4e02723a2f9bfa89"} Jan 21 12:07:54 crc kubenswrapper[4824]: I0121 12:07:54.441519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerStarted","Data":"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a"} Jan 21 12:07:55 crc kubenswrapper[4824]: I0121 12:07:55.459083 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerID="23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a" exitCode=0 Jan 21 12:07:55 crc kubenswrapper[4824]: I0121 12:07:55.459282 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerDied","Data":"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a"} Jan 21 12:07:56 crc kubenswrapper[4824]: I0121 12:07:56.470320 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerStarted","Data":"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd"} Jan 21 12:07:56 crc kubenswrapper[4824]: I0121 12:07:56.496205 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vznz5" podStartSLOduration=1.980979414 podStartE2EDuration="4.496188228s" podCreationTimestamp="2026-01-21 12:07:52 +0000 UTC" firstStartedPulling="2026-01-21 12:07:53.42980227 +0000 UTC m=+3415.722831562" lastFinishedPulling="2026-01-21 12:07:55.945011085 +0000 UTC m=+3418.238040376" observedRunningTime="2026-01-21 12:07:56.490314173 +0000 UTC m=+3418.783343465" watchObservedRunningTime="2026-01-21 12:07:56.496188228 +0000 UTC m=+3418.789217520" Jan 21 12:08:01 crc kubenswrapper[4824]: I0121 12:08:01.406456 4824 scope.go:117] "RemoveContainer" containerID="c6793d88c2985e5e3d28f1236121b047a23b50c741fdeb44a8e3cc182cf06683" Jan 21 12:08:01 crc kubenswrapper[4824]: I0121 12:08:01.433743 4824 scope.go:117] "RemoveContainer" containerID="6e4471e366c5d457dbdb31676ebe2e10113437d05155002808726047c10699c0" Jan 21 12:08:02 crc kubenswrapper[4824]: I0121 12:08:02.471781 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:02 crc kubenswrapper[4824]: I0121 12:08:02.472094 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:02 crc kubenswrapper[4824]: I0121 12:08:02.509612 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:02 crc kubenswrapper[4824]: I0121 12:08:02.575568 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:02 crc kubenswrapper[4824]: I0121 12:08:02.747611 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:08:04 crc kubenswrapper[4824]: I0121 12:08:04.560205 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vznz5" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="registry-server" containerID="cri-o://b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd" gracePeriod=2 Jan 21 12:08:04 crc kubenswrapper[4824]: I0121 12:08:04.962220 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.063324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities\") pod \"7d34ec22-f51b-4776-8342-80b16e3c41e0\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.064124 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities" (OuterVolumeSpecName: "utilities") pod "7d34ec22-f51b-4776-8342-80b16e3c41e0" (UID: "7d34ec22-f51b-4776-8342-80b16e3c41e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.082199 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.183240 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf6dz\" (UniqueName: \"kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz\") pod \"7d34ec22-f51b-4776-8342-80b16e3c41e0\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.183388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content\") pod \"7d34ec22-f51b-4776-8342-80b16e3c41e0\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.193114 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz" (OuterVolumeSpecName: "kube-api-access-wf6dz") pod "7d34ec22-f51b-4776-8342-80b16e3c41e0" (UID: "7d34ec22-f51b-4776-8342-80b16e3c41e0"). InnerVolumeSpecName "kube-api-access-wf6dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.230409 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d34ec22-f51b-4776-8342-80b16e3c41e0" (UID: "7d34ec22-f51b-4776-8342-80b16e3c41e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.288004 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf6dz\" (UniqueName: \"kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz\") on node \"crc\" DevicePath \"\"" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.288039 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.573793 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerID="b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd" exitCode=0 Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.573885 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerDied","Data":"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd"} Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.573925 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.573967 4824 scope.go:117] "RemoveContainer" containerID="b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.573936 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerDied","Data":"736f0aba8b28cb81d6e6ab52fa2cf14cf8d30b070cc2c19d4e02723a2f9bfa89"} Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.594453 4824 scope.go:117] "RemoveContainer" containerID="23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.608229 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.615436 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.626129 4824 scope.go:117] "RemoveContainer" containerID="5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.646463 4824 scope.go:117] "RemoveContainer" containerID="b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd" Jan 21 12:08:05 crc kubenswrapper[4824]: E0121 12:08:05.646879 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd\": container with ID starting with b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd not found: ID does not exist" containerID="b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.646928 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd"} err="failed to get container status \"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd\": rpc error: code = NotFound desc = could not find container \"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd\": container with ID starting with b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd not found: ID does not exist" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.646973 4824 scope.go:117] "RemoveContainer" containerID="23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a" Jan 21 12:08:05 crc kubenswrapper[4824]: E0121 12:08:05.647284 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a\": container with ID starting with 23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a not found: ID does not exist" containerID="23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.647317 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a"} err="failed to get container status \"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a\": rpc error: code = NotFound desc = could not find container \"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a\": container with ID starting with 23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a not found: ID does not exist" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.647343 4824 scope.go:117] "RemoveContainer" containerID="5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59" Jan 21 12:08:05 crc kubenswrapper[4824]: E0121 12:08:05.647836 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59\": container with ID starting with 5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59 not found: ID does not exist" containerID="5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.647864 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59"} err="failed to get container status \"5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59\": rpc error: code = NotFound desc = could not find container \"5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59\": container with ID starting with 5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59 not found: ID does not exist" Jan 21 12:08:06 crc kubenswrapper[4824]: I0121 12:08:06.061846 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" path="/var/lib/kubelet/pods/7d34ec22-f51b-4776-8342-80b16e3c41e0/volumes" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.644891 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:08:57 crc kubenswrapper[4824]: E0121 12:08:57.645730 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="extract-content" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.645746 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="extract-content" Jan 21 12:08:57 crc kubenswrapper[4824]: E0121 12:08:57.645767 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="extract-utilities" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.645774 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="extract-utilities" Jan 21 12:08:57 crc kubenswrapper[4824]: E0121 12:08:57.645784 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="registry-server" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.645790 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="registry-server" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.646010 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="registry-server" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.647292 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.663315 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.692043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgzqh\" (UniqueName: \"kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.692285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.692505 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.794480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgzqh\" (UniqueName: \"kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.794596 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.794731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.795260 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.795280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.811194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgzqh\" (UniqueName: \"kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.969315 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:58 crc kubenswrapper[4824]: I0121 12:08:58.445561 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.122104 4824 generic.go:334] "Generic (PLEG): container finished" podID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerID="f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91" exitCode=0 Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.122217 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerDied","Data":"f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91"} Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.122498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerStarted","Data":"1d304bfa846d8ce7ddcda5a95d370eb81f9bbae489b9d7c8996d46294146e282"} Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.125162 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.125234 4824 generic.go:334] "Generic (PLEG): container finished" podID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerID="a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe" exitCode=0 Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.125293 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/must-gather-z5tfn" event={"ID":"2245f836-eda4-4bf5-aa8a-5b18de2927b2","Type":"ContainerDied","Data":"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe"} Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.125835 4824 scope.go:117] "RemoveContainer" containerID="a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe" Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.664643 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-87fjw_must-gather-z5tfn_2245f836-eda4-4bf5-aa8a-5b18de2927b2/gather/0.log" Jan 21 12:09:00 crc kubenswrapper[4824]: I0121 12:09:00.151989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerStarted","Data":"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb"} Jan 21 12:09:01 crc kubenswrapper[4824]: I0121 12:09:01.162680 4824 generic.go:334] "Generic (PLEG): container finished" podID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerID="9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb" exitCode=0 Jan 21 12:09:01 crc kubenswrapper[4824]: I0121 12:09:01.162776 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerDied","Data":"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb"} Jan 21 12:09:02 crc kubenswrapper[4824]: I0121 12:09:02.173028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerStarted","Data":"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce"} Jan 21 12:09:02 crc kubenswrapper[4824]: I0121 12:09:02.196525 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-45vjg" podStartSLOduration=2.6812069149999997 podStartE2EDuration="5.196506392s" podCreationTimestamp="2026-01-21 12:08:57 +0000 UTC" firstStartedPulling="2026-01-21 12:08:59.124322213 +0000 UTC m=+3481.417351494" lastFinishedPulling="2026-01-21 12:09:01.639621679 +0000 UTC m=+3483.932650971" observedRunningTime="2026-01-21 12:09:02.187758589 +0000 UTC m=+3484.480787882" watchObservedRunningTime="2026-01-21 12:09:02.196506392 +0000 UTC m=+3484.489535684" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.332025 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-87fjw/must-gather-z5tfn"] Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.333144 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-87fjw/must-gather-z5tfn" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="copy" containerID="cri-o://a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320" gracePeriod=2 Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.342583 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-87fjw/must-gather-z5tfn"] Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.743685 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-87fjw_must-gather-z5tfn_2245f836-eda4-4bf5-aa8a-5b18de2927b2/copy/0.log" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.744413 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.759925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjtv2\" (UniqueName: \"kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2\") pod \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.760280 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output\") pod \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.765744 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2" (OuterVolumeSpecName: "kube-api-access-qjtv2") pod "2245f836-eda4-4bf5-aa8a-5b18de2927b2" (UID: "2245f836-eda4-4bf5-aa8a-5b18de2927b2"). InnerVolumeSpecName "kube-api-access-qjtv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.862626 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjtv2\" (UniqueName: \"kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2\") on node \"crc\" DevicePath \"\"" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.912950 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2245f836-eda4-4bf5-aa8a-5b18de2927b2" (UID: "2245f836-eda4-4bf5-aa8a-5b18de2927b2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.966757 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.969500 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.970152 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.015071 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.059376 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" path="/var/lib/kubelet/pods/2245f836-eda4-4bf5-aa8a-5b18de2927b2/volumes" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.224882 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-87fjw_must-gather-z5tfn_2245f836-eda4-4bf5-aa8a-5b18de2927b2/copy/0.log" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.225212 4824 generic.go:334] "Generic (PLEG): container finished" podID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerID="a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320" exitCode=143 Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.225933 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.226315 4824 scope.go:117] "RemoveContainer" containerID="a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.252555 4824 scope.go:117] "RemoveContainer" containerID="a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.269752 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.315277 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.315372 4824 scope.go:117] "RemoveContainer" containerID="a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320" Jan 21 12:09:08 crc kubenswrapper[4824]: E0121 12:09:08.315784 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320\": container with ID starting with a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320 not found: ID does not exist" containerID="a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.315813 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320"} err="failed to get container status \"a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320\": rpc error: code = NotFound desc = could not find container \"a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320\": container with ID starting with a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320 not found: ID does not exist" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.315833 4824 scope.go:117] "RemoveContainer" containerID="a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe" Jan 21 12:09:08 crc kubenswrapper[4824]: E0121 12:09:08.316305 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe\": container with ID starting with a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe not found: ID does not exist" containerID="a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.316332 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe"} err="failed to get container status \"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe\": rpc error: code = NotFound desc = could not find container \"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe\": container with ID starting with a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe not found: ID does not exist" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.247896 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-45vjg" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="registry-server" containerID="cri-o://703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce" gracePeriod=2 Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.620809 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.722415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities\") pod \"c37315cd-e029-42bf-bfeb-4e5c497833e4\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.722659 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgzqh\" (UniqueName: \"kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh\") pod \"c37315cd-e029-42bf-bfeb-4e5c497833e4\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.722735 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content\") pod \"c37315cd-e029-42bf-bfeb-4e5c497833e4\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.723202 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities" (OuterVolumeSpecName: "utilities") pod "c37315cd-e029-42bf-bfeb-4e5c497833e4" (UID: "c37315cd-e029-42bf-bfeb-4e5c497833e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.728098 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh" (OuterVolumeSpecName: "kube-api-access-rgzqh") pod "c37315cd-e029-42bf-bfeb-4e5c497833e4" (UID: "c37315cd-e029-42bf-bfeb-4e5c497833e4"). InnerVolumeSpecName "kube-api-access-rgzqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.740607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c37315cd-e029-42bf-bfeb-4e5c497833e4" (UID: "c37315cd-e029-42bf-bfeb-4e5c497833e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.826184 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.826591 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.826607 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgzqh\" (UniqueName: \"kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh\") on node \"crc\" DevicePath \"\"" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.259362 4824 generic.go:334] "Generic (PLEG): container finished" podID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerID="703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce" exitCode=0 Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.259428 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerDied","Data":"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce"} Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.259476 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerDied","Data":"1d304bfa846d8ce7ddcda5a95d370eb81f9bbae489b9d7c8996d46294146e282"} Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.259478 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.259500 4824 scope.go:117] "RemoveContainer" containerID="703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.291635 4824 scope.go:117] "RemoveContainer" containerID="9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.295732 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.314492 4824 scope.go:117] "RemoveContainer" containerID="f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.317826 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.353215 4824 scope.go:117] "RemoveContainer" containerID="703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce" Jan 21 12:09:11 crc kubenswrapper[4824]: E0121 12:09:11.353723 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce\": container with ID starting with 703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce not found: ID does not exist" containerID="703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.353770 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce"} err="failed to get container status \"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce\": rpc error: code = NotFound desc = could not find container \"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce\": container with ID starting with 703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce not found: ID does not exist" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.353802 4824 scope.go:117] "RemoveContainer" containerID="9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb" Jan 21 12:09:11 crc kubenswrapper[4824]: E0121 12:09:11.354370 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb\": container with ID starting with 9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb not found: ID does not exist" containerID="9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.354415 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb"} err="failed to get container status \"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb\": rpc error: code = NotFound desc = could not find container \"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb\": container with ID starting with 9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb not found: ID does not exist" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.354454 4824 scope.go:117] "RemoveContainer" containerID="f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91" Jan 21 12:09:11 crc kubenswrapper[4824]: E0121 12:09:11.354771 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91\": container with ID starting with f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91 not found: ID does not exist" containerID="f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.354797 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91"} err="failed to get container status \"f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91\": rpc error: code = NotFound desc = could not find container \"f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91\": container with ID starting with f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91 not found: ID does not exist" Jan 21 12:09:12 crc kubenswrapper[4824]: I0121 12:09:12.059714 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" path="/var/lib/kubelet/pods/c37315cd-e029-42bf-bfeb-4e5c497833e4/volumes" Jan 21 12:09:46 crc kubenswrapper[4824]: I0121 12:09:46.064975 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:09:46 crc kubenswrapper[4824]: I0121 12:09:46.065694 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:10:16 crc kubenswrapper[4824]: I0121 12:10:16.065768 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:10:16 crc kubenswrapper[4824]: I0121 12:10:16.066461 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.101234 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5vql/must-gather-jkhmz"] Jan 21 12:10:35 crc kubenswrapper[4824]: E0121 12:10:35.102236 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="gather" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102251 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="gather" Jan 21 12:10:35 crc kubenswrapper[4824]: E0121 12:10:35.102261 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="registry-server" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102267 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="registry-server" Jan 21 12:10:35 crc kubenswrapper[4824]: E0121 12:10:35.102278 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="copy" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102284 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="copy" Jan 21 12:10:35 crc kubenswrapper[4824]: E0121 12:10:35.102297 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="extract-utilities" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102303 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="extract-utilities" Jan 21 12:10:35 crc kubenswrapper[4824]: E0121 12:10:35.102316 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="extract-content" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102321 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="extract-content" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102511 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="registry-server" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102535 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="copy" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102544 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="gather" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.103536 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.105433 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n5vql"/"openshift-service-ca.crt" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.105686 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-n5vql"/"default-dockercfg-2zcwc" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.105815 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n5vql"/"kube-root-ca.crt" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.119290 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n5vql/must-gather-jkhmz"] Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.170267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdgjw\" (UniqueName: \"kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.170378 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.272892 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.273133 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdgjw\" (UniqueName: \"kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.273384 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.290707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdgjw\" (UniqueName: \"kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.426776 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.837233 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n5vql/must-gather-jkhmz"] Jan 21 12:10:36 crc kubenswrapper[4824]: I0121 12:10:36.040611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/must-gather-jkhmz" event={"ID":"e1c80411-6f38-43b2-8656-639978632741","Type":"ContainerStarted","Data":"65ed37028631bc528541d931c3708c861369ab00cb45d0d01989f2e346501aa0"} Jan 21 12:10:37 crc kubenswrapper[4824]: I0121 12:10:37.052566 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/must-gather-jkhmz" event={"ID":"e1c80411-6f38-43b2-8656-639978632741","Type":"ContainerStarted","Data":"d0b75fbd713ed74fa81844f4c6ccb4b25946f8456765e9861e694df4f9f52b15"} Jan 21 12:10:37 crc kubenswrapper[4824]: I0121 12:10:37.052991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/must-gather-jkhmz" event={"ID":"e1c80411-6f38-43b2-8656-639978632741","Type":"ContainerStarted","Data":"0937d1d47c67d983a6ee92637b82dba42aee39048f14d63afc09202819ad4750"} Jan 21 12:10:37 crc kubenswrapper[4824]: I0121 12:10:37.072979 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n5vql/must-gather-jkhmz" podStartSLOduration=2.072946813 podStartE2EDuration="2.072946813s" podCreationTimestamp="2026-01-21 12:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 12:10:37.070699456 +0000 UTC m=+3579.363728748" watchObservedRunningTime="2026-01-21 12:10:37.072946813 +0000 UTC m=+3579.365976095" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.572088 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5vql/crc-debug-9n2zc"] Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.573826 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.765131 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rbp4\" (UniqueName: \"kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.765335 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.868151 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rbp4\" (UniqueName: \"kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.868469 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.868604 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.888455 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rbp4\" (UniqueName: \"kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.893829 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: W0121 12:10:39.924735 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2b727cf_5339_4da5_be88_b46b2bcf5dd1.slice/crio-2730b37ace706deb236dc24c607e69f539af5cbe65b803a5f377d1ca122f5af3 WatchSource:0}: Error finding container 2730b37ace706deb236dc24c607e69f539af5cbe65b803a5f377d1ca122f5af3: Status 404 returned error can't find the container with id 2730b37ace706deb236dc24c607e69f539af5cbe65b803a5f377d1ca122f5af3 Jan 21 12:10:40 crc kubenswrapper[4824]: I0121 12:10:40.099909 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" event={"ID":"d2b727cf-5339-4da5-be88-b46b2bcf5dd1","Type":"ContainerStarted","Data":"2730b37ace706deb236dc24c607e69f539af5cbe65b803a5f377d1ca122f5af3"} Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.109269 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" event={"ID":"d2b727cf-5339-4da5-be88-b46b2bcf5dd1","Type":"ContainerStarted","Data":"2719d0b32f33e24c79ae317ca212bc0f5429a69ed705aafd8970f2d93f3e60b9"} Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.128647 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" podStartSLOduration=2.128633542 podStartE2EDuration="2.128633542s" podCreationTimestamp="2026-01-21 12:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 12:10:41.120545735 +0000 UTC m=+3583.413575026" watchObservedRunningTime="2026-01-21 12:10:41.128633542 +0000 UTC m=+3583.421662834" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.892470 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7f965fb4-8slr2_3f7f97b5-8abd-4351-8396-cf1b76c376a7/barbican-api-log/0.log" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.903888 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7f965fb4-8slr2_3f7f97b5-8abd-4351-8396-cf1b76c376a7/barbican-api/0.log" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.960221 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b7cfb8cd-ktpp5_8b84edbe-4485-4027-bfc4-cebbb1b10543/barbican-keystone-listener-log/0.log" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.966424 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b7cfb8cd-ktpp5_8b84edbe-4485-4027-bfc4-cebbb1b10543/barbican-keystone-listener/0.log" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.982738 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7cf777545f-f8drb_283e6982-7444-48c7-b7ac-ed4210329a08/barbican-worker-log/0.log" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.989709 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7cf777545f-f8drb_283e6982-7444-48c7-b7ac-ed4210329a08/barbican-worker/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.025571 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6_300a037e-ac25-4a9e-9339-6cc707969f7c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.052187 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/ceilometer-central-agent/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.074700 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/ceilometer-notification-agent/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.078458 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/sg-core/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.088883 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/proxy-httpd/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.105023 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ff9de537-ecd2-469d-b90e-5857bbf64c2d/cinder-api-log/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.155316 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ff9de537-ecd2-469d-b90e-5857bbf64c2d/cinder-api/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.202215 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e0f0267-58d2-4690-9187-8ddf03ef082e/cinder-scheduler/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.231228 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e0f0267-58d2-4690-9187-8ddf03ef082e/probe/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.251133 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz_b8c7ec7b-ee3f-46bc-95f8-4406713ba43a/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.269972 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n_8d46d1eb-4230-49c7-8376-698e7689ad2b/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.315367 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-26d8b_976a6446-e79c-42db-be45-43147767f1da/dnsmasq-dns/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.318472 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-26d8b_976a6446-e79c-42db-be45-43147767f1da/init/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.338536 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9_8dbaf72e-72a7-4073-a4c1-95cbdabaa64b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.348643 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_219725d0-9bef-4e42-8b78-cbc74a8ac9db/glance-log/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.369743 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_219725d0-9bef-4e42-8b78-cbc74a8ac9db/glance-httpd/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.385868 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7419ae4a-e8ea-4384-a9b8-3a3e02e27f25/glance-log/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.407747 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7419ae4a-e8ea-4384-a9b8-3a3e02e27f25/glance-httpd/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.844183 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-549445f5fc-fwxlz_8655cad5-2680-4cf6-b50a-0745adeeb469/heat-api/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.127934 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-68b8cf894c-ws5zd_a1aefb21-aac8-4f8f-a402-16006c1f336a/heat-cfnapi/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.232428 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5bc6fc99dd-nv45v_ca8530d7-bcfb-45fa-b0ff-a2702c081bda/heat-engine/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.248420 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2_b821ea8d-9327-4623-b5be-3da7d3872cab/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.273435 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-b74gh_60fedbf7-07e6-41be-ada3-2a06f28835e3/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.638773 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c5fc769c9-spmzr_7d943b5e-d05f-445a-888d-7ebe340468b9/keystone-api/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.651348 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29483281-bz8hk_c92d6f21-1fe2-41a2-a516-a0d0740ab36f/keystone-cron/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.662187 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_d42233a9-cfd9-486d-b314-f17d6a9d3b03/kube-state-metrics/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.685084 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn_fdde8991-c204-45a0-b344-dcdc41a9a275/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:46 crc kubenswrapper[4824]: I0121 12:10:46.065105 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:10:46 crc kubenswrapper[4824]: I0121 12:10:46.065623 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:10:46 crc kubenswrapper[4824]: I0121 12:10:46.070590 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 12:10:46 crc kubenswrapper[4824]: I0121 12:10:46.071026 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 12:10:46 crc kubenswrapper[4824]: I0121 12:10:46.071091 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" gracePeriod=600 Jan 21 12:10:46 crc kubenswrapper[4824]: E0121 12:10:46.191379 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:10:47 crc kubenswrapper[4824]: I0121 12:10:47.160097 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" exitCode=0 Jan 21 12:10:47 crc kubenswrapper[4824]: I0121 12:10:47.160166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4"} Jan 21 12:10:47 crc kubenswrapper[4824]: I0121 12:10:47.160379 4824 scope.go:117] "RemoveContainer" containerID="253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d" Jan 21 12:10:47 crc kubenswrapper[4824]: I0121 12:10:47.161311 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:10:47 crc kubenswrapper[4824]: E0121 12:10:47.161675 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.050267 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:11:00 crc kubenswrapper[4824]: E0121 12:11:00.050750 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.070115 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_dba0050b-8a73-4355-a1e0-7c9a03557ead/memcached/0.log" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.171266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-659565f76f-jdg2q_b6f2ba03-746e-465c-bde2-c917129fb125/neutron-api/0.log" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.229881 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-659565f76f-jdg2q_b6f2ba03-746e-465c-bde2-c917129fb125/neutron-httpd/0.log" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.251307 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l_4f3c3a58-5d6b-4cc2-9707-d088b2fd463d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.523095 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55525bd3-ae2b-494d-9a33-dd7d00c576b2/nova-api-log/0.log" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.980640 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55525bd3-ae2b-494d-9a33-dd7d00c576b2/nova-api-api/0.log" Jan 21 12:11:01 crc kubenswrapper[4824]: I0121 12:11:01.122349 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8effdacf-8fec-40b4-8f61-8856ac0f232c/nova-cell0-conductor-conductor/0.log" Jan 21 12:11:01 crc kubenswrapper[4824]: I0121 12:11:01.243525 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c398d1e9-13e1-400c-a609-bceea49aea51/nova-cell1-conductor-conductor/0.log" Jan 21 12:11:01 crc kubenswrapper[4824]: I0121 12:11:01.348506 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8e42e9d6-485f-4b97-b049-316cd47af30e/nova-cell1-novncproxy-novncproxy/0.log" Jan 21 12:11:01 crc kubenswrapper[4824]: I0121 12:11:01.402490 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-7crw8_7355a8fc-8c18-4de1-8baf-f6294ea9e11c/nova-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:01 crc kubenswrapper[4824]: I0121 12:11:01.497599 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_654a41d0-b44a-4194-a514-a3c60126d0d9/nova-metadata-log/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.522445 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_654a41d0-b44a-4194-a514-a3c60126d0d9/nova-metadata-metadata/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.678159 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_7860ffbb-b435-4736-b2d3-f8871e70dc7a/nova-scheduler-scheduler/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.697342 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cd8d62b-ab74-4a1b-95f5-7253ed6c0346/galera/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.720168 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cd8d62b-ab74-4a1b-95f5-7253ed6c0346/mysql-bootstrap/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.901714 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_218aca21-4c37-4862-b891-fc228baa5348/galera/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.911937 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_218aca21-4c37-4862-b891-fc228baa5348/mysql-bootstrap/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.931199 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1af94d08-6d1e-4a97-8f82-27a1755d17a1/openstackclient/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.946428 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-htwd6_511f2b6d-a08d-49f8-b393-ab222219d4a7/ovn-controller/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.953140 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vrpmw_8652f3f6-1ad7-49d9-9048-827be17a3dd0/openstack-network-exporter/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.963285 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovsdb-server/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.972175 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovs-vswitchd/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.976717 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovsdb-server-init/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.005683 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gf4zq_18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.016373 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b9c25b1-d0b9-430e-88cf-0aeaf6785461/ovn-northd/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.022359 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b9c25b1-d0b9-430e-88cf-0aeaf6785461/openstack-network-exporter/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.040695 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5c0b5876-da04-49e6-8946-4fc6060a64ec/ovsdbserver-nb/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.046109 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5c0b5876-da04-49e6-8946-4fc6060a64ec/openstack-network-exporter/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.063301 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef/ovsdbserver-sb/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.067562 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef/openstack-network-exporter/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.154323 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8654bbf998-p8crm_c23477ba-5a41-44da-a8b7-6606595b9726/placement-log/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.223685 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8654bbf998-p8crm_c23477ba-5a41-44da-a8b7-6606595b9726/placement-api/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.244086 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d8e256a-aa1b-4ccb-8e8c-c86be1e31762/rabbitmq/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.247409 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d8e256a-aa1b-4ccb-8e8c-c86be1e31762/setup-container/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.270850 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4a4d725-1a49-4292-af74-6ab09ab15cdb/rabbitmq/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.274912 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4a4d725-1a49-4292-af74-6ab09ab15cdb/setup-container/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.289701 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w_890acdab-9f31-4afc-80e4-c1df308ac5a6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.297327 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-2flzn_34c68ecf-cf29-44cb-9880-a3d45f625454/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.307345 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf_e3664d03-d53d-4a49-8cab-c4266f2d1426/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.317256 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pn4sw_ad533443-01b7-4860-9920-93ed67f6b52f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.332490 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-wbrwb_4529344f-3dfd-4858-83f1-4abffe47f2ad/ssh-known-hosts-edpm-deployment/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.474432 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-79f4c948b7-4hz4p_01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d/proxy-httpd/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.504393 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-79f4c948b7-4hz4p_01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d/proxy-server/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.515608 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-nc6vf_29473359-3517-4b86-bd5e-80e25706ff27/swift-ring-rebalance/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.559703 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-server/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.574383 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.580147 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.588466 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-replicator/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.593038 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-auditor/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.602458 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-reaper/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.603945 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.613727 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-server/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.647457 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-replicator/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.652840 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-auditor/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.664293 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-updater/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.673724 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-server/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.703695 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-replicator/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.731306 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-auditor/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.737626 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-updater/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.750929 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-expirer/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.766525 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/rsync/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.781445 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/swift-recon-cron/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.887614 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6_49b2a136-e5c3-4741-80c6-4c545d4a82a0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.942686 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_58d30951-ee56-4217-a614-a04951369c8b/tempest-tests-tempest-tests-runner/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.949836 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_333160e2-3c05-401f-af89-abae8c0cb9d9/test-operator-logs-container/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.985546 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz_d672f153-6ef2-432d-829a-db63c8daab69/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.009262 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.064031 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.075317 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.090178 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.095270 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.108383 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.229048 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.312063 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.327462 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.328276 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.332999 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.339746 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.345668 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.348525 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.352989 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.359627 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.370672 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.377752 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.410781 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.424049 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.731450 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.742041 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.844471 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.853095 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.855698 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.861531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.915924 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.949132 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:11:06 crc kubenswrapper[4824]: I0121 12:11:06.016192 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:11:06 crc kubenswrapper[4824]: I0121 12:11:06.024441 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:11:06 crc kubenswrapper[4824]: I0121 12:11:06.037536 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:11:06 crc kubenswrapper[4824]: I0121 12:11:06.172216 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.109619 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.190804 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.238773 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.256473 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.277528 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.296351 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.354529 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.361411 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.371077 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:11:11 crc kubenswrapper[4824]: I0121 12:11:11.368776 4824 generic.go:334] "Generic (PLEG): container finished" podID="d2b727cf-5339-4da5-be88-b46b2bcf5dd1" containerID="2719d0b32f33e24c79ae317ca212bc0f5429a69ed705aafd8970f2d93f3e60b9" exitCode=0 Jan 21 12:11:11 crc kubenswrapper[4824]: I0121 12:11:11.368867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" event={"ID":"d2b727cf-5339-4da5-be88-b46b2bcf5dd1","Type":"ContainerDied","Data":"2719d0b32f33e24c79ae317ca212bc0f5429a69ed705aafd8970f2d93f3e60b9"} Jan 21 12:11:11 crc kubenswrapper[4824]: I0121 12:11:11.981353 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-r65dm_c117483a-fe5c-423b-995e-9c2e7f59566a/control-plane-machine-set-operator/0.log" Jan 21 12:11:11 crc kubenswrapper[4824]: I0121 12:11:11.992643 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/kube-rbac-proxy/0.log" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.000481 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/machine-api-operator/0.log" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.455482 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.486294 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-9n2zc"] Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.492035 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-9n2zc"] Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.541549 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rbp4\" (UniqueName: \"kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4\") pod \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.542008 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host\") pod \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.542143 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host" (OuterVolumeSpecName: "host") pod "d2b727cf-5339-4da5-be88-b46b2bcf5dd1" (UID: "d2b727cf-5339-4da5-be88-b46b2bcf5dd1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.542753 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.547442 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4" (OuterVolumeSpecName: "kube-api-access-9rbp4") pod "d2b727cf-5339-4da5-be88-b46b2bcf5dd1" (UID: "d2b727cf-5339-4da5-be88-b46b2bcf5dd1"). InnerVolumeSpecName "kube-api-access-9rbp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.644474 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rbp4\" (UniqueName: \"kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.389059 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2730b37ace706deb236dc24c607e69f539af5cbe65b803a5f377d1ca122f5af3" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.389151 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.620404 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5vql/crc-debug-twlgv"] Jan 21 12:11:13 crc kubenswrapper[4824]: E0121 12:11:13.621579 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2b727cf-5339-4da5-be88-b46b2bcf5dd1" containerName="container-00" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.621676 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2b727cf-5339-4da5-be88-b46b2bcf5dd1" containerName="container-00" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.621983 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2b727cf-5339-4da5-be88-b46b2bcf5dd1" containerName="container-00" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.623059 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.667095 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdz7h\" (UniqueName: \"kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.667170 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.768698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdz7h\" (UniqueName: \"kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.768766 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.768992 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.795279 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdz7h\" (UniqueName: \"kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.942035 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:14 crc kubenswrapper[4824]: I0121 12:11:14.058694 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2b727cf-5339-4da5-be88-b46b2bcf5dd1" path="/var/lib/kubelet/pods/d2b727cf-5339-4da5-be88-b46b2bcf5dd1/volumes" Jan 21 12:11:14 crc kubenswrapper[4824]: I0121 12:11:14.401158 4824 generic.go:334] "Generic (PLEG): container finished" podID="fd685561-3fbc-4a45-85aa-4acfdd089880" containerID="07bf87c3a875d48665a8353210d054b6f5a80d43f1b2712cff93bd33cf233d5b" exitCode=0 Jan 21 12:11:14 crc kubenswrapper[4824]: I0121 12:11:14.401249 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-twlgv" event={"ID":"fd685561-3fbc-4a45-85aa-4acfdd089880","Type":"ContainerDied","Data":"07bf87c3a875d48665a8353210d054b6f5a80d43f1b2712cff93bd33cf233d5b"} Jan 21 12:11:14 crc kubenswrapper[4824]: I0121 12:11:14.401336 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-twlgv" event={"ID":"fd685561-3fbc-4a45-85aa-4acfdd089880","Type":"ContainerStarted","Data":"d22ae759971e42ba5d4c5d7745735291590008b6acb73204a7c4d5bd657d6073"} Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.057539 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:11:15 crc kubenswrapper[4824]: E0121 12:11:15.058931 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.484262 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.506115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host\") pod \"fd685561-3fbc-4a45-85aa-4acfdd089880\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.506193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdz7h\" (UniqueName: \"kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h\") pod \"fd685561-3fbc-4a45-85aa-4acfdd089880\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.509518 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host" (OuterVolumeSpecName: "host") pod "fd685561-3fbc-4a45-85aa-4acfdd089880" (UID: "fd685561-3fbc-4a45-85aa-4acfdd089880"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.514296 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h" (OuterVolumeSpecName: "kube-api-access-zdz7h") pod "fd685561-3fbc-4a45-85aa-4acfdd089880" (UID: "fd685561-3fbc-4a45-85aa-4acfdd089880"). InnerVolumeSpecName "kube-api-access-zdz7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.608243 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.608277 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdz7h\" (UniqueName: \"kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:16 crc kubenswrapper[4824]: I0121 12:11:16.199472 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-twlgv"] Jan 21 12:11:16 crc kubenswrapper[4824]: I0121 12:11:16.206344 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-twlgv"] Jan 21 12:11:16 crc kubenswrapper[4824]: I0121 12:11:16.419497 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d22ae759971e42ba5d4c5d7745735291590008b6acb73204a7c4d5bd657d6073" Jan 21 12:11:16 crc kubenswrapper[4824]: I0121 12:11:16.419565 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.329762 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5vql/crc-debug-vzzhf"] Jan 21 12:11:17 crc kubenswrapper[4824]: E0121 12:11:17.330447 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd685561-3fbc-4a45-85aa-4acfdd089880" containerName="container-00" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.330462 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd685561-3fbc-4a45-85aa-4acfdd089880" containerName="container-00" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.330636 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd685561-3fbc-4a45-85aa-4acfdd089880" containerName="container-00" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.331225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.456472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.456580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zbzk\" (UniqueName: \"kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.557748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.557950 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.558087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zbzk\" (UniqueName: \"kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.577510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zbzk\" (UniqueName: \"kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.647765 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.061941 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd685561-3fbc-4a45-85aa-4acfdd089880" path="/var/lib/kubelet/pods/fd685561-3fbc-4a45-85aa-4acfdd089880/volumes" Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.450485 4824 generic.go:334] "Generic (PLEG): container finished" podID="f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" containerID="28a96abcbc61a2ef1455e05edd37bad363143f763a99e610aa47efbbc2d11cc9" exitCode=0 Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.450577 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" event={"ID":"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4","Type":"ContainerDied","Data":"28a96abcbc61a2ef1455e05edd37bad363143f763a99e610aa47efbbc2d11cc9"} Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.451042 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" event={"ID":"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4","Type":"ContainerStarted","Data":"feaebf64732df5c5fc007c0740c2d69db7d3fdf4dcdabb5f6f984ad0740b5eae"} Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.490768 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-vzzhf"] Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.498648 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-vzzhf"] Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.972175 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.986371 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.996710 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.549833 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.709423 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host\") pod \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.709532 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zbzk\" (UniqueName: \"kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk\") pod \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.709546 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host" (OuterVolumeSpecName: "host") pod "f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" (UID: "f96e30dd-f9f8-4f4f-9dee-27b8275b17c4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.711054 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.723012 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk" (OuterVolumeSpecName: "kube-api-access-7zbzk") pod "f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" (UID: "f96e30dd-f9f8-4f4f-9dee-27b8275b17c4"). InnerVolumeSpecName "kube-api-access-7zbzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.813303 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zbzk\" (UniqueName: \"kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:20 crc kubenswrapper[4824]: I0121 12:11:20.058233 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" path="/var/lib/kubelet/pods/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4/volumes" Jan 21 12:11:20 crc kubenswrapper[4824]: I0121 12:11:20.473216 4824 scope.go:117] "RemoveContainer" containerID="28a96abcbc61a2ef1455e05edd37bad363143f763a99e610aa47efbbc2d11cc9" Jan 21 12:11:20 crc kubenswrapper[4824]: I0121 12:11:20.473256 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.687036 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-xjqs5_1575ab9d-016f-403a-990b-3b4c4d53f6f1/nmstate-console-plugin/0.log" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.707484 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qjpv9_a01ecff0-9051-4ddb-8c8e-a4f379b6c823/nmstate-handler/0.log" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.716586 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/nmstate-metrics/0.log" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.724266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/kube-rbac-proxy/0.log" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.734280 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-zmc4k_df60ca53-0573-4a6a-9235-86e0a1c17f91/nmstate-operator/0.log" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.745680 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-h6kkh_edf54909-b53e-44de-90bc-355e23bbc767/nmstate-webhook/0.log" Jan 21 12:11:27 crc kubenswrapper[4824]: I0121 12:11:27.049942 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:11:27 crc kubenswrapper[4824]: E0121 12:11:27.050754 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:11:33 crc kubenswrapper[4824]: I0121 12:11:33.414382 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:11:33 crc kubenswrapper[4824]: I0121 12:11:33.420870 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:11:33 crc kubenswrapper[4824]: I0121 12:11:33.435584 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.846277 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.858697 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.863192 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.871832 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.878861 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.885819 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.892994 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.900115 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.909707 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.934678 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.943524 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:11:35 crc kubenswrapper[4824]: I0121 12:11:35.312284 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:11:35 crc kubenswrapper[4824]: I0121 12:11:35.321390 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.320775 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/extract/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.328337 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/util/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.335782 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/pull/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.343532 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/extract/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.349821 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/util/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.355868 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/pull/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.781238 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/registry-server/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.788528 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/extract-utilities/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.793306 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/extract-content/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.318548 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/registry-server/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.323196 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/extract-utilities/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.329719 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/extract-content/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.342113 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xcxs9_b4c29122-d9b2-4f85-8b29-c0e7d888667f/marketplace-operator/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.493118 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/registry-server/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.497301 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/extract-utilities/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.502769 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/extract-content/0.log" Jan 21 12:11:40 crc kubenswrapper[4824]: I0121 12:11:40.049080 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:11:40 crc kubenswrapper[4824]: E0121 12:11:40.049752 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:11:40 crc kubenswrapper[4824]: I0121 12:11:40.064615 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/registry-server/0.log" Jan 21 12:11:40 crc kubenswrapper[4824]: I0121 12:11:40.071905 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/extract-utilities/0.log" Jan 21 12:11:40 crc kubenswrapper[4824]: I0121 12:11:40.077211 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/extract-content/0.log" Jan 21 12:11:53 crc kubenswrapper[4824]: I0121 12:11:53.049300 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:11:53 crc kubenswrapper[4824]: E0121 12:11:53.051144 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:12:04 crc kubenswrapper[4824]: I0121 12:12:04.049759 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:12:04 crc kubenswrapper[4824]: E0121 12:12:04.050854 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:12:16 crc kubenswrapper[4824]: I0121 12:12:16.050206 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:12:16 crc kubenswrapper[4824]: E0121 12:12:16.051091 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:12:29 crc kubenswrapper[4824]: I0121 12:12:29.049593 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:12:29 crc kubenswrapper[4824]: E0121 12:12:29.050330 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:12:37 crc kubenswrapper[4824]: I0121 12:12:37.942266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:12:37 crc kubenswrapper[4824]: I0121 12:12:37.949894 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:12:37 crc kubenswrapper[4824]: I0121 12:12:37.977796 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:12:37 crc kubenswrapper[4824]: I0121 12:12:37.999266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:12:38 crc kubenswrapper[4824]: I0121 12:12:38.017230 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:12:38 crc kubenswrapper[4824]: I0121 12:12:38.034784 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:12:38 crc kubenswrapper[4824]: I0121 12:12:38.929005 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:12:38 crc kubenswrapper[4824]: I0121 12:12:38.981668 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:12:38 crc kubenswrapper[4824]: I0121 12:12:38.997901 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.011190 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.024025 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.035187 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.147922 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.235050 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.244558 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.563479 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.574073 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.579345 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.588829 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.595141 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.601802 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.608900 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.616063 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.624051 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.626261 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.635421 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.652276 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.660732 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.719452 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.731144 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.771684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.829244 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.949745 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.961674 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.987431 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:12:40 crc kubenswrapper[4824]: I0121 12:12:40.152774 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:12:40 crc kubenswrapper[4824]: I0121 12:12:40.163085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:12:40 crc kubenswrapper[4824]: I0121 12:12:40.201718 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.050375 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.065354 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.072590 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.353728 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.362654 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.407209 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.435510 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.453700 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.473920 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.538914 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.549701 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.558341 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.706589 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-r65dm_c117483a-fe5c-423b-995e-9c2e7f59566a/control-plane-machine-set-operator/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.715792 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/kube-rbac-proxy/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.727399 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/machine-api-operator/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.049261 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:12:42 crc kubenswrapper[4824]: E0121 12:12:42.049570 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.413669 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.454811 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.466339 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.471416 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.480094 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.490477 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.559884 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.637522 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.654103 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.803688 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-xjqs5_1575ab9d-016f-403a-990b-3b4c4d53f6f1/nmstate-console-plugin/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.822439 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qjpv9_a01ecff0-9051-4ddb-8c8e-a4f379b6c823/nmstate-handler/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.831287 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/nmstate-metrics/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.838407 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/kube-rbac-proxy/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.859324 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-zmc4k_df60ca53-0573-4a6a-9235-86e0a1c17f91/nmstate-operator/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.869038 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-h6kkh_edf54909-b53e-44de-90bc-355e23bbc767/nmstate-webhook/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.949553 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.960860 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.026354 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.036389 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.066486 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.116159 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.192998 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.205177 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.221212 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.379845 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.700728 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.714464 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.770602 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.792141 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.821878 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.848202 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.912827 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.922344 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.933588 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.345904 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/kube-multus-additional-cni-plugins/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.353029 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/egress-router-binary-copy/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.359544 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/cni-plugins/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.365359 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/bond-cni-plugin/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.371022 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/routeoverride-cni/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.376541 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/whereabouts-cni-bincopy/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.381813 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/whereabouts-cni/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.412004 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-27d2q_7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34/multus-admission-controller/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.415708 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-27d2q_7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34/kube-rbac-proxy/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.456032 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/2.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.518411 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/3.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.548755 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-9rmnw_8f9bd12b-22ee-44ed-a8b9-c100d043f691/network-metrics-daemon/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.553537 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-9rmnw_8f9bd12b-22ee-44ed-a8b9-c100d043f691/kube-rbac-proxy/0.log" Jan 21 12:12:53 crc kubenswrapper[4824]: I0121 12:12:53.049347 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:12:53 crc kubenswrapper[4824]: E0121 12:12:53.050046 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:13:04 crc kubenswrapper[4824]: I0121 12:13:04.049529 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:13:04 crc kubenswrapper[4824]: E0121 12:13:04.050505 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:13:19 crc kubenswrapper[4824]: I0121 12:13:19.051249 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:13:19 crc kubenswrapper[4824]: E0121 12:13:19.052240 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:13:32 crc kubenswrapper[4824]: I0121 12:13:32.049826 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:13:32 crc kubenswrapper[4824]: E0121 12:13:32.051092 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:13:45 crc kubenswrapper[4824]: I0121 12:13:45.051374 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:13:45 crc kubenswrapper[4824]: E0121 12:13:45.052398 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:13:56 crc kubenswrapper[4824]: I0121 12:13:56.049738 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:13:56 crc kubenswrapper[4824]: E0121 12:13:56.050594 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:08 crc kubenswrapper[4824]: I0121 12:14:08.059986 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:14:08 crc kubenswrapper[4824]: E0121 12:14:08.060898 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:20 crc kubenswrapper[4824]: I0121 12:14:20.050323 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:14:20 crc kubenswrapper[4824]: E0121 12:14:20.051171 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.050255 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:14:33 crc kubenswrapper[4824]: E0121 12:14:33.051137 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.760157 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:33 crc kubenswrapper[4824]: E0121 12:14:33.760866 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" containerName="container-00" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.760978 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" containerName="container-00" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.761249 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" containerName="container-00" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.762763 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.768191 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.815845 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgk5l\" (UniqueName: \"kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.815895 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.815945 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.918640 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.919177 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.919357 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.919653 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.919869 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgk5l\" (UniqueName: \"kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.944008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgk5l\" (UniqueName: \"kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:34 crc kubenswrapper[4824]: I0121 12:14:34.094488 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:34 crc kubenswrapper[4824]: I0121 12:14:34.524029 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:35 crc kubenswrapper[4824]: I0121 12:14:35.170470 4824 generic.go:334] "Generic (PLEG): container finished" podID="38d18f54-9d85-4302-81d2-f678960544f3" containerID="bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494" exitCode=0 Jan 21 12:14:35 crc kubenswrapper[4824]: I0121 12:14:35.170593 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerDied","Data":"bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494"} Jan 21 12:14:35 crc kubenswrapper[4824]: I0121 12:14:35.170890 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerStarted","Data":"85ab2a5e6d6c105b66b6695758ecd564e9e8e6bb0a07a420c6427da89d4eadb4"} Jan 21 12:14:35 crc kubenswrapper[4824]: I0121 12:14:35.172622 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 12:14:36 crc kubenswrapper[4824]: I0121 12:14:36.183149 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerStarted","Data":"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c"} Jan 21 12:14:37 crc kubenswrapper[4824]: I0121 12:14:37.193433 4824 generic.go:334] "Generic (PLEG): container finished" podID="38d18f54-9d85-4302-81d2-f678960544f3" containerID="07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c" exitCode=0 Jan 21 12:14:37 crc kubenswrapper[4824]: I0121 12:14:37.193557 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerDied","Data":"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c"} Jan 21 12:14:38 crc kubenswrapper[4824]: I0121 12:14:38.207222 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerStarted","Data":"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2"} Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.094614 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.095441 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.139498 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.160071 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8ctsm" podStartSLOduration=8.633838012 podStartE2EDuration="11.160034758s" podCreationTimestamp="2026-01-21 12:14:33 +0000 UTC" firstStartedPulling="2026-01-21 12:14:35.172243026 +0000 UTC m=+3817.465272318" lastFinishedPulling="2026-01-21 12:14:37.698439782 +0000 UTC m=+3819.991469064" observedRunningTime="2026-01-21 12:14:38.231483261 +0000 UTC m=+3820.524512553" watchObservedRunningTime="2026-01-21 12:14:44.160034758 +0000 UTC m=+3826.453064050" Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.337822 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.385906 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.050684 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:14:46 crc kubenswrapper[4824]: E0121 12:14:46.051433 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.326069 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8ctsm" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="registry-server" containerID="cri-o://a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2" gracePeriod=2 Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.722293 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.766684 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities\") pod \"38d18f54-9d85-4302-81d2-f678960544f3\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.766748 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content\") pod \"38d18f54-9d85-4302-81d2-f678960544f3\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.767410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities" (OuterVolumeSpecName: "utilities") pod "38d18f54-9d85-4302-81d2-f678960544f3" (UID: "38d18f54-9d85-4302-81d2-f678960544f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.805462 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38d18f54-9d85-4302-81d2-f678960544f3" (UID: "38d18f54-9d85-4302-81d2-f678960544f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.868232 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgk5l\" (UniqueName: \"kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l\") pod \"38d18f54-9d85-4302-81d2-f678960544f3\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.869360 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.869391 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.873699 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l" (OuterVolumeSpecName: "kube-api-access-jgk5l") pod "38d18f54-9d85-4302-81d2-f678960544f3" (UID: "38d18f54-9d85-4302-81d2-f678960544f3"). InnerVolumeSpecName "kube-api-access-jgk5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.971598 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgk5l\" (UniqueName: \"kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l\") on node \"crc\" DevicePath \"\"" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.336848 4824 generic.go:334] "Generic (PLEG): container finished" podID="38d18f54-9d85-4302-81d2-f678960544f3" containerID="a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2" exitCode=0 Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.337092 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerDied","Data":"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2"} Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.337856 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerDied","Data":"85ab2a5e6d6c105b66b6695758ecd564e9e8e6bb0a07a420c6427da89d4eadb4"} Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.337918 4824 scope.go:117] "RemoveContainer" containerID="a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.337228 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.360207 4824 scope.go:117] "RemoveContainer" containerID="07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.379892 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.386377 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.400838 4824 scope.go:117] "RemoveContainer" containerID="bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.422674 4824 scope.go:117] "RemoveContainer" containerID="a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2" Jan 21 12:14:47 crc kubenswrapper[4824]: E0121 12:14:47.423243 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2\": container with ID starting with a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2 not found: ID does not exist" containerID="a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.423279 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2"} err="failed to get container status \"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2\": rpc error: code = NotFound desc = could not find container \"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2\": container with ID starting with a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2 not found: ID does not exist" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.423304 4824 scope.go:117] "RemoveContainer" containerID="07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c" Jan 21 12:14:47 crc kubenswrapper[4824]: E0121 12:14:47.423760 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c\": container with ID starting with 07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c not found: ID does not exist" containerID="07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.423862 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c"} err="failed to get container status \"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c\": rpc error: code = NotFound desc = could not find container \"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c\": container with ID starting with 07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c not found: ID does not exist" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.423950 4824 scope.go:117] "RemoveContainer" containerID="bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494" Jan 21 12:14:47 crc kubenswrapper[4824]: E0121 12:14:47.424548 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494\": container with ID starting with bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494 not found: ID does not exist" containerID="bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.424581 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494"} err="failed to get container status \"bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494\": rpc error: code = NotFound desc = could not find container \"bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494\": container with ID starting with bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494 not found: ID does not exist" Jan 21 12:14:48 crc kubenswrapper[4824]: I0121 12:14:48.059341 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38d18f54-9d85-4302-81d2-f678960544f3" path="/var/lib/kubelet/pods/38d18f54-9d85-4302-81d2-f678960544f3/volumes" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.192012 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:14:58 crc kubenswrapper[4824]: E0121 12:14:58.192893 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="extract-utilities" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.192908 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="extract-utilities" Jan 21 12:14:58 crc kubenswrapper[4824]: E0121 12:14:58.192946 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="extract-content" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.192974 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="extract-content" Jan 21 12:14:58 crc kubenswrapper[4824]: E0121 12:14:58.192986 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="registry-server" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.192993 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="registry-server" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.193207 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="registry-server" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.194461 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.200259 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.370407 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmcqm\" (UniqueName: \"kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.370476 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.370526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.472926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmcqm\" (UniqueName: \"kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.473009 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.473057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.473566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.473801 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.492384 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmcqm\" (UniqueName: \"kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.522084 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.986997 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:14:59 crc kubenswrapper[4824]: I0121 12:14:59.050893 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:14:59 crc kubenswrapper[4824]: E0121 12:14:59.051235 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:59 crc kubenswrapper[4824]: I0121 12:14:59.456361 4824 generic.go:334] "Generic (PLEG): container finished" podID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerID="fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170" exitCode=0 Jan 21 12:14:59 crc kubenswrapper[4824]: I0121 12:14:59.456426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerDied","Data":"fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170"} Jan 21 12:14:59 crc kubenswrapper[4824]: I0121 12:14:59.456471 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerStarted","Data":"ba64d2b0b833265f8588d182cbd463930b1169951658d8b2c78ac3122d41d2dc"} Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.174522 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69"] Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.176928 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.184569 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.184713 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.184952 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69"] Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.217502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.217574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv4dm\" (UniqueName: \"kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.217620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.320313 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.320398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv4dm\" (UniqueName: \"kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.320431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.321400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.327612 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.334951 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv4dm\" (UniqueName: \"kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.496622 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.970906 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69"] Jan 21 12:15:01 crc kubenswrapper[4824]: I0121 12:15:01.476475 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerStarted","Data":"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444"} Jan 21 12:15:01 crc kubenswrapper[4824]: I0121 12:15:01.479319 4824 generic.go:334] "Generic (PLEG): container finished" podID="81672912-3e15-467d-9c3c-8a6a3b297d28" containerID="f4ebcbc8adfa426ef5d7d6d30a337211dc204ce7b6a3c0e1a84921e00e5e354f" exitCode=0 Jan 21 12:15:01 crc kubenswrapper[4824]: I0121 12:15:01.479354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" event={"ID":"81672912-3e15-467d-9c3c-8a6a3b297d28","Type":"ContainerDied","Data":"f4ebcbc8adfa426ef5d7d6d30a337211dc204ce7b6a3c0e1a84921e00e5e354f"} Jan 21 12:15:01 crc kubenswrapper[4824]: I0121 12:15:01.479372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" event={"ID":"81672912-3e15-467d-9c3c-8a6a3b297d28","Type":"ContainerStarted","Data":"31b5621ec04585ca9ce7815dbcbeb50db0208149c8f301f56e27d700b8ba7e23"} Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.779120 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.973049 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume\") pod \"81672912-3e15-467d-9c3c-8a6a3b297d28\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.973175 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv4dm\" (UniqueName: \"kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm\") pod \"81672912-3e15-467d-9c3c-8a6a3b297d28\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.973260 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume\") pod \"81672912-3e15-467d-9c3c-8a6a3b297d28\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.974087 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume" (OuterVolumeSpecName: "config-volume") pod "81672912-3e15-467d-9c3c-8a6a3b297d28" (UID: "81672912-3e15-467d-9c3c-8a6a3b297d28"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.985073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "81672912-3e15-467d-9c3c-8a6a3b297d28" (UID: "81672912-3e15-467d-9c3c-8a6a3b297d28"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.989840 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm" (OuterVolumeSpecName: "kube-api-access-hv4dm") pod "81672912-3e15-467d-9c3c-8a6a3b297d28" (UID: "81672912-3e15-467d-9c3c-8a6a3b297d28"). InnerVolumeSpecName "kube-api-access-hv4dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.076176 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.076215 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv4dm\" (UniqueName: \"kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.076229 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.500988 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" event={"ID":"81672912-3e15-467d-9c3c-8a6a3b297d28","Type":"ContainerDied","Data":"31b5621ec04585ca9ce7815dbcbeb50db0208149c8f301f56e27d700b8ba7e23"} Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.501043 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.501049 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31b5621ec04585ca9ce7815dbcbeb50db0208149c8f301f56e27d700b8ba7e23" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.862786 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz"] Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.867847 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz"] Jan 21 12:15:04 crc kubenswrapper[4824]: I0121 12:15:04.060637 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c02bf47-f856-44f3-af2c-cc96c45471c1" path="/var/lib/kubelet/pods/4c02bf47-f856-44f3-af2c-cc96c45471c1/volumes" Jan 21 12:15:04 crc kubenswrapper[4824]: I0121 12:15:04.521516 4824 generic.go:334] "Generic (PLEG): container finished" podID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerID="b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444" exitCode=0 Jan 21 12:15:04 crc kubenswrapper[4824]: I0121 12:15:04.521568 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerDied","Data":"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444"} Jan 21 12:15:05 crc kubenswrapper[4824]: I0121 12:15:05.532815 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerStarted","Data":"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25"} Jan 21 12:15:05 crc kubenswrapper[4824]: I0121 12:15:05.570627 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lbnms" podStartSLOduration=1.997598907 podStartE2EDuration="7.570605504s" podCreationTimestamp="2026-01-21 12:14:58 +0000 UTC" firstStartedPulling="2026-01-21 12:14:59.458580143 +0000 UTC m=+3841.751609435" lastFinishedPulling="2026-01-21 12:15:05.03158674 +0000 UTC m=+3847.324616032" observedRunningTime="2026-01-21 12:15:05.56509791 +0000 UTC m=+3847.858127202" watchObservedRunningTime="2026-01-21 12:15:05.570605504 +0000 UTC m=+3847.863634796" Jan 21 12:15:08 crc kubenswrapper[4824]: I0121 12:15:08.523804 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:08 crc kubenswrapper[4824]: I0121 12:15:08.524556 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:09 crc kubenswrapper[4824]: I0121 12:15:09.563451 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lbnms" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="registry-server" probeResult="failure" output=< Jan 21 12:15:09 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Jan 21 12:15:09 crc kubenswrapper[4824]: > Jan 21 12:15:10 crc kubenswrapper[4824]: I0121 12:15:10.050544 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:15:10 crc kubenswrapper[4824]: E0121 12:15:10.052047 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:15:18 crc kubenswrapper[4824]: I0121 12:15:18.570099 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:18 crc kubenswrapper[4824]: I0121 12:15:18.624662 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:18 crc kubenswrapper[4824]: I0121 12:15:18.806820 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:15:19 crc kubenswrapper[4824]: I0121 12:15:19.679257 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lbnms" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="registry-server" containerID="cri-o://c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25" gracePeriod=2 Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.128177 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.187197 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities\") pod \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.187306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmcqm\" (UniqueName: \"kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm\") pod \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.187489 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content\") pod \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.188727 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities" (OuterVolumeSpecName: "utilities") pod "5c25db90-0383-4d60-bc0d-e5fdd37540cd" (UID: "5c25db90-0383-4d60-bc0d-e5fdd37540cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.195881 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm" (OuterVolumeSpecName: "kube-api-access-dmcqm") pod "5c25db90-0383-4d60-bc0d-e5fdd37540cd" (UID: "5c25db90-0383-4d60-bc0d-e5fdd37540cd"). InnerVolumeSpecName "kube-api-access-dmcqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.289153 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.289184 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmcqm\" (UniqueName: \"kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.289762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c25db90-0383-4d60-bc0d-e5fdd37540cd" (UID: "5c25db90-0383-4d60-bc0d-e5fdd37540cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.390419 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.690397 4824 generic.go:334] "Generic (PLEG): container finished" podID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerID="c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25" exitCode=0 Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.690450 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerDied","Data":"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25"} Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.690497 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerDied","Data":"ba64d2b0b833265f8588d182cbd463930b1169951658d8b2c78ac3122d41d2dc"} Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.690520 4824 scope.go:117] "RemoveContainer" containerID="c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.690687 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.715127 4824 scope.go:117] "RemoveContainer" containerID="b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.730603 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.739010 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.752852 4824 scope.go:117] "RemoveContainer" containerID="fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.775291 4824 scope.go:117] "RemoveContainer" containerID="c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25" Jan 21 12:15:20 crc kubenswrapper[4824]: E0121 12:15:20.775663 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25\": container with ID starting with c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25 not found: ID does not exist" containerID="c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.775699 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25"} err="failed to get container status \"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25\": rpc error: code = NotFound desc = could not find container \"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25\": container with ID starting with c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25 not found: ID does not exist" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.775725 4824 scope.go:117] "RemoveContainer" containerID="b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444" Jan 21 12:15:20 crc kubenswrapper[4824]: E0121 12:15:20.776011 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444\": container with ID starting with b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444 not found: ID does not exist" containerID="b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.776037 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444"} err="failed to get container status \"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444\": rpc error: code = NotFound desc = could not find container \"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444\": container with ID starting with b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444 not found: ID does not exist" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.776053 4824 scope.go:117] "RemoveContainer" containerID="fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170" Jan 21 12:15:20 crc kubenswrapper[4824]: E0121 12:15:20.776348 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170\": container with ID starting with fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170 not found: ID does not exist" containerID="fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.776369 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170"} err="failed to get container status \"fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170\": rpc error: code = NotFound desc = could not find container \"fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170\": container with ID starting with fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170 not found: ID does not exist" Jan 21 12:15:22 crc kubenswrapper[4824]: I0121 12:15:22.059477 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" path="/var/lib/kubelet/pods/5c25db90-0383-4d60-bc0d-e5fdd37540cd/volumes" Jan 21 12:15:24 crc kubenswrapper[4824]: I0121 12:15:24.049166 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:15:24 crc kubenswrapper[4824]: E0121 12:15:24.049887 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:15:38 crc kubenswrapper[4824]: I0121 12:15:38.056095 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:15:38 crc kubenswrapper[4824]: E0121 12:15:38.057184 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:15:53 crc kubenswrapper[4824]: I0121 12:15:53.050054 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:15:53 crc kubenswrapper[4824]: I0121 12:15:53.993229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"d9d0949df611c52cc67603e5984be93524636cdbd8451c97adc3963dfdbe83d5"} Jan 21 12:15:55 crc kubenswrapper[4824]: I0121 12:15:55.970994 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-79f4c948b7-4hz4p" podUID="01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Jan 21 12:16:01 crc kubenswrapper[4824]: I0121 12:16:01.696948 4824 scope.go:117] "RemoveContainer" containerID="dd264d8d4b02d26df3192da84e03b87f54958338166fb00a9d9562f0fbc98220" Jan 21 12:17:01 crc kubenswrapper[4824]: I0121 12:17:01.779861 4824 scope.go:117] "RemoveContainer" containerID="2719d0b32f33e24c79ae317ca212bc0f5429a69ed705aafd8970f2d93f3e60b9" Jan 21 12:18:01 crc kubenswrapper[4824]: I0121 12:18:01.841664 4824 scope.go:117] "RemoveContainer" containerID="07bf87c3a875d48665a8353210d054b6f5a80d43f1b2712cff93bd33cf233d5b" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.749356 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:15 crc kubenswrapper[4824]: E0121 12:18:15.750256 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="extract-content" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750273 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="extract-content" Jan 21 12:18:15 crc kubenswrapper[4824]: E0121 12:18:15.750291 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="registry-server" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750297 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="registry-server" Jan 21 12:18:15 crc kubenswrapper[4824]: E0121 12:18:15.750308 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81672912-3e15-467d-9c3c-8a6a3b297d28" containerName="collect-profiles" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750315 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="81672912-3e15-467d-9c3c-8a6a3b297d28" containerName="collect-profiles" Jan 21 12:18:15 crc kubenswrapper[4824]: E0121 12:18:15.750345 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="extract-utilities" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750351 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="extract-utilities" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750569 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="81672912-3e15-467d-9c3c-8a6a3b297d28" containerName="collect-profiles" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750584 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="registry-server" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.751851 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.757247 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.922177 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.922324 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.922893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nmcb\" (UniqueName: \"kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.025712 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.025775 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.025858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nmcb\" (UniqueName: \"kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.026193 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.026420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.049239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nmcb\" (UniqueName: \"kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.065747 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.065805 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.075841 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.552257 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:17 crc kubenswrapper[4824]: I0121 12:18:17.335772 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerID="1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5" exitCode=0 Jan 21 12:18:17 crc kubenswrapper[4824]: I0121 12:18:17.335885 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerDied","Data":"1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5"} Jan 21 12:18:17 crc kubenswrapper[4824]: I0121 12:18:17.336117 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerStarted","Data":"b9354b8fbffbcb7440fe899058d36ff7c2aea9e5cfda3cc98ecfdcc6a57082cf"} Jan 21 12:18:18 crc kubenswrapper[4824]: I0121 12:18:18.348736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerStarted","Data":"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819"} Jan 21 12:18:19 crc kubenswrapper[4824]: I0121 12:18:19.358562 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerID="3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819" exitCode=0 Jan 21 12:18:19 crc kubenswrapper[4824]: I0121 12:18:19.358917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerDied","Data":"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819"} Jan 21 12:18:20 crc kubenswrapper[4824]: I0121 12:18:20.369827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerStarted","Data":"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40"} Jan 21 12:18:20 crc kubenswrapper[4824]: I0121 12:18:20.386129 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-prbq4" podStartSLOduration=2.7814604149999997 podStartE2EDuration="5.386115407s" podCreationTimestamp="2026-01-21 12:18:15 +0000 UTC" firstStartedPulling="2026-01-21 12:18:17.337829693 +0000 UTC m=+4039.630858975" lastFinishedPulling="2026-01-21 12:18:19.942484675 +0000 UTC m=+4042.235513967" observedRunningTime="2026-01-21 12:18:20.383575799 +0000 UTC m=+4042.676605092" watchObservedRunningTime="2026-01-21 12:18:20.386115407 +0000 UTC m=+4042.679144699" Jan 21 12:18:26 crc kubenswrapper[4824]: I0121 12:18:26.076866 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:26 crc kubenswrapper[4824]: I0121 12:18:26.077581 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:26 crc kubenswrapper[4824]: I0121 12:18:26.121754 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:26 crc kubenswrapper[4824]: I0121 12:18:26.458092 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:26 crc kubenswrapper[4824]: I0121 12:18:26.496563 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:28 crc kubenswrapper[4824]: I0121 12:18:28.434175 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-prbq4" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="registry-server" containerID="cri-o://b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40" gracePeriod=2 Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.355811 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.444890 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerID="b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40" exitCode=0 Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.444943 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerDied","Data":"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40"} Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.444989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerDied","Data":"b9354b8fbffbcb7440fe899058d36ff7c2aea9e5cfda3cc98ecfdcc6a57082cf"} Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.445010 4824 scope.go:117] "RemoveContainer" containerID="b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.445025 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.475539 4824 scope.go:117] "RemoveContainer" containerID="3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.500739 4824 scope.go:117] "RemoveContainer" containerID="1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.510472 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content\") pod \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.510777 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nmcb\" (UniqueName: \"kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb\") pod \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.510989 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities\") pod \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.511598 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities" (OuterVolumeSpecName: "utilities") pod "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" (UID: "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.511786 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.519236 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb" (OuterVolumeSpecName: "kube-api-access-7nmcb") pod "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" (UID: "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1"). InnerVolumeSpecName "kube-api-access-7nmcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.556858 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" (UID: "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.568090 4824 scope.go:117] "RemoveContainer" containerID="b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40" Jan 21 12:18:29 crc kubenswrapper[4824]: E0121 12:18:29.568511 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40\": container with ID starting with b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40 not found: ID does not exist" containerID="b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.568550 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40"} err="failed to get container status \"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40\": rpc error: code = NotFound desc = could not find container \"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40\": container with ID starting with b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40 not found: ID does not exist" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.568574 4824 scope.go:117] "RemoveContainer" containerID="3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819" Jan 21 12:18:29 crc kubenswrapper[4824]: E0121 12:18:29.568945 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819\": container with ID starting with 3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819 not found: ID does not exist" containerID="3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.569009 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819"} err="failed to get container status \"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819\": rpc error: code = NotFound desc = could not find container \"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819\": container with ID starting with 3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819 not found: ID does not exist" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.569048 4824 scope.go:117] "RemoveContainer" containerID="1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5" Jan 21 12:18:29 crc kubenswrapper[4824]: E0121 12:18:29.569358 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5\": container with ID starting with 1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5 not found: ID does not exist" containerID="1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.569390 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5"} err="failed to get container status \"1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5\": rpc error: code = NotFound desc = could not find container \"1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5\": container with ID starting with 1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5 not found: ID does not exist" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.614042 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.614076 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nmcb\" (UniqueName: \"kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb\") on node \"crc\" DevicePath \"\"" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.780621 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.793542 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:30 crc kubenswrapper[4824]: I0121 12:18:30.066191 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" path="/var/lib/kubelet/pods/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1/volumes" Jan 21 12:18:46 crc kubenswrapper[4824]: I0121 12:18:46.065133 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:18:46 crc kubenswrapper[4824]: I0121 12:18:46.066047 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:18:51 crc kubenswrapper[4824]: I0121 12:18:51.634651 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1c80411-6f38-43b2-8656-639978632741" containerID="d0b75fbd713ed74fa81844f4c6ccb4b25946f8456765e9861e694df4f9f52b15" exitCode=0 Jan 21 12:18:51 crc kubenswrapper[4824]: I0121 12:18:51.634741 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/must-gather-jkhmz" event={"ID":"e1c80411-6f38-43b2-8656-639978632741","Type":"ContainerDied","Data":"d0b75fbd713ed74fa81844f4c6ccb4b25946f8456765e9861e694df4f9f52b15"} Jan 21 12:18:51 crc kubenswrapper[4824]: I0121 12:18:51.635618 4824 scope.go:117] "RemoveContainer" containerID="d0b75fbd713ed74fa81844f4c6ccb4b25946f8456765e9861e694df4f9f52b15" Jan 21 12:18:51 crc kubenswrapper[4824]: I0121 12:18:51.694737 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n5vql_must-gather-jkhmz_e1c80411-6f38-43b2-8656-639978632741/gather/0.log" Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.325166 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5vql/must-gather-jkhmz"] Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.326124 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-n5vql/must-gather-jkhmz" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="copy" containerID="cri-o://0937d1d47c67d983a6ee92637b82dba42aee39048f14d63afc09202819ad4750" gracePeriod=2 Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.331113 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5vql/must-gather-jkhmz"] Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.720514 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n5vql_must-gather-jkhmz_e1c80411-6f38-43b2-8656-639978632741/copy/0.log" Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.721105 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1c80411-6f38-43b2-8656-639978632741" containerID="0937d1d47c67d983a6ee92637b82dba42aee39048f14d63afc09202819ad4750" exitCode=143 Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.721157 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65ed37028631bc528541d931c3708c861369ab00cb45d0d01989f2e346501aa0" Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.728518 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n5vql_must-gather-jkhmz_e1c80411-6f38-43b2-8656-639978632741/copy/0.log" Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.728927 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.914473 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdgjw\" (UniqueName: \"kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw\") pod \"e1c80411-6f38-43b2-8656-639978632741\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.914833 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output\") pod \"e1c80411-6f38-43b2-8656-639978632741\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.927122 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw" (OuterVolumeSpecName: "kube-api-access-fdgjw") pod "e1c80411-6f38-43b2-8656-639978632741" (UID: "e1c80411-6f38-43b2-8656-639978632741"). InnerVolumeSpecName "kube-api-access-fdgjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:19:01 crc kubenswrapper[4824]: I0121 12:19:01.016525 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdgjw\" (UniqueName: \"kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw\") on node \"crc\" DevicePath \"\"" Jan 21 12:19:01 crc kubenswrapper[4824]: I0121 12:19:01.096332 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e1c80411-6f38-43b2-8656-639978632741" (UID: "e1c80411-6f38-43b2-8656-639978632741"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:19:01 crc kubenswrapper[4824]: I0121 12:19:01.117645 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 21 12:19:01 crc kubenswrapper[4824]: I0121 12:19:01.730155 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:19:02 crc kubenswrapper[4824]: I0121 12:19:02.064331 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1c80411-6f38-43b2-8656-639978632741" path="/var/lib/kubelet/pods/e1c80411-6f38-43b2-8656-639978632741/volumes" Jan 21 12:19:02 crc kubenswrapper[4824]: I0121 12:19:02.237846 4824 scope.go:117] "RemoveContainer" containerID="0937d1d47c67d983a6ee92637b82dba42aee39048f14d63afc09202819ad4750" Jan 21 12:19:02 crc kubenswrapper[4824]: I0121 12:19:02.277373 4824 scope.go:117] "RemoveContainer" containerID="d0b75fbd713ed74fa81844f4c6ccb4b25946f8456765e9861e694df4f9f52b15" Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.065249 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.065881 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.065931 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.066396 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9d0949df611c52cc67603e5984be93524636cdbd8451c97adc3963dfdbe83d5"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.066450 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://d9d0949df611c52cc67603e5984be93524636cdbd8451c97adc3963dfdbe83d5" gracePeriod=600 Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.858202 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="d9d0949df611c52cc67603e5984be93524636cdbd8451c97adc3963dfdbe83d5" exitCode=0 Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.858278 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"d9d0949df611c52cc67603e5984be93524636cdbd8451c97adc3963dfdbe83d5"} Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.859009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"111c22954de19114358173702fa304b6c5ccb7d35d262c1f76a3e710a19b138b"} Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.859044 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.556867 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:19:56 crc kubenswrapper[4824]: E0121 12:19:56.557837 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="extract-content" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.557851 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="extract-content" Jan 21 12:19:56 crc kubenswrapper[4824]: E0121 12:19:56.557866 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="registry-server" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.557871 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="registry-server" Jan 21 12:19:56 crc kubenswrapper[4824]: E0121 12:19:56.557885 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="copy" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.557891 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="copy" Jan 21 12:19:56 crc kubenswrapper[4824]: E0121 12:19:56.557909 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="extract-utilities" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.557915 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="extract-utilities" Jan 21 12:19:56 crc kubenswrapper[4824]: E0121 12:19:56.557927 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="gather" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.557932 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="gather" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.558113 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="registry-server" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.558130 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="copy" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.558149 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="gather" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.559438 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.566773 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.623837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-976n5\" (UniqueName: \"kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.624032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.624207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.725786 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.726034 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-976n5\" (UniqueName: \"kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.726117 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.726392 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.726526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.744061 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-976n5\" (UniqueName: \"kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.880477 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:57 crc kubenswrapper[4824]: I0121 12:19:57.340525 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:19:58 crc kubenswrapper[4824]: I0121 12:19:58.199779 4824 generic.go:334] "Generic (PLEG): container finished" podID="10d70708-c370-4167-be39-ef73bf918cce" containerID="254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d" exitCode=0 Jan 21 12:19:58 crc kubenswrapper[4824]: I0121 12:19:58.199826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerDied","Data":"254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d"} Jan 21 12:19:58 crc kubenswrapper[4824]: I0121 12:19:58.200197 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerStarted","Data":"aa80bc3d83a5ca946ba8cf648ddeffd27293e1a25cfe3b133babcd1feb735afd"} Jan 21 12:19:58 crc kubenswrapper[4824]: I0121 12:19:58.202014 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 12:19:59 crc kubenswrapper[4824]: I0121 12:19:59.211913 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerStarted","Data":"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221"} Jan 21 12:20:00 crc kubenswrapper[4824]: I0121 12:20:00.223174 4824 generic.go:334] "Generic (PLEG): container finished" podID="10d70708-c370-4167-be39-ef73bf918cce" containerID="9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221" exitCode=0 Jan 21 12:20:00 crc kubenswrapper[4824]: I0121 12:20:00.223219 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerDied","Data":"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221"} Jan 21 12:20:01 crc kubenswrapper[4824]: I0121 12:20:01.234702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerStarted","Data":"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d"} Jan 21 12:20:01 crc kubenswrapper[4824]: I0121 12:20:01.263617 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5449q" podStartSLOduration=2.7529481369999997 podStartE2EDuration="5.263599853s" podCreationTimestamp="2026-01-21 12:19:56 +0000 UTC" firstStartedPulling="2026-01-21 12:19:58.201734575 +0000 UTC m=+4140.494763867" lastFinishedPulling="2026-01-21 12:20:00.71238629 +0000 UTC m=+4143.005415583" observedRunningTime="2026-01-21 12:20:01.258015533 +0000 UTC m=+4143.551044825" watchObservedRunningTime="2026-01-21 12:20:01.263599853 +0000 UTC m=+4143.556629145" Jan 21 12:20:06 crc kubenswrapper[4824]: I0121 12:20:06.882010 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:06 crc kubenswrapper[4824]: I0121 12:20:06.882760 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:06 crc kubenswrapper[4824]: I0121 12:20:06.919066 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:07 crc kubenswrapper[4824]: I0121 12:20:07.331697 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:07 crc kubenswrapper[4824]: I0121 12:20:07.373562 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.311011 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5449q" podUID="10d70708-c370-4167-be39-ef73bf918cce" containerName="registry-server" containerID="cri-o://8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d" gracePeriod=2 Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.766013 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.795368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities\") pod \"10d70708-c370-4167-be39-ef73bf918cce\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.796161 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities" (OuterVolumeSpecName: "utilities") pod "10d70708-c370-4167-be39-ef73bf918cce" (UID: "10d70708-c370-4167-be39-ef73bf918cce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.796620 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.898935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content\") pod \"10d70708-c370-4167-be39-ef73bf918cce\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.899033 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-976n5\" (UniqueName: \"kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5\") pod \"10d70708-c370-4167-be39-ef73bf918cce\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.906309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5" (OuterVolumeSpecName: "kube-api-access-976n5") pod "10d70708-c370-4167-be39-ef73bf918cce" (UID: "10d70708-c370-4167-be39-ef73bf918cce"). InnerVolumeSpecName "kube-api-access-976n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.921991 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10d70708-c370-4167-be39-ef73bf918cce" (UID: "10d70708-c370-4167-be39-ef73bf918cce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.001829 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.001877 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-976n5\" (UniqueName: \"kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5\") on node \"crc\" DevicePath \"\"" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.320205 4824 generic.go:334] "Generic (PLEG): container finished" podID="10d70708-c370-4167-be39-ef73bf918cce" containerID="8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d" exitCode=0 Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.320252 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerDied","Data":"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d"} Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.320259 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.320286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerDied","Data":"aa80bc3d83a5ca946ba8cf648ddeffd27293e1a25cfe3b133babcd1feb735afd"} Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.320308 4824 scope.go:117] "RemoveContainer" containerID="8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.338392 4824 scope.go:117] "RemoveContainer" containerID="9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.339429 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.344694 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.357437 4824 scope.go:117] "RemoveContainer" containerID="254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.410465 4824 scope.go:117] "RemoveContainer" containerID="8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d" Jan 21 12:20:10 crc kubenswrapper[4824]: E0121 12:20:10.410976 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d\": container with ID starting with 8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d not found: ID does not exist" containerID="8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.411032 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d"} err="failed to get container status \"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d\": rpc error: code = NotFound desc = could not find container \"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d\": container with ID starting with 8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d not found: ID does not exist" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.411065 4824 scope.go:117] "RemoveContainer" containerID="9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221" Jan 21 12:20:10 crc kubenswrapper[4824]: E0121 12:20:10.411494 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221\": container with ID starting with 9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221 not found: ID does not exist" containerID="9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.411536 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221"} err="failed to get container status \"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221\": rpc error: code = NotFound desc = could not find container \"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221\": container with ID starting with 9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221 not found: ID does not exist" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.411578 4824 scope.go:117] "RemoveContainer" containerID="254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d" Jan 21 12:20:10 crc kubenswrapper[4824]: E0121 12:20:10.411850 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d\": container with ID starting with 254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d not found: ID does not exist" containerID="254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.411884 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d"} err="failed to get container status \"254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d\": rpc error: code = NotFound desc = could not find container \"254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d\": container with ID starting with 254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d not found: ID does not exist" Jan 21 12:20:12 crc kubenswrapper[4824]: I0121 12:20:12.060949 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10d70708-c370-4167-be39-ef73bf918cce" path="/var/lib/kubelet/pods/10d70708-c370-4167-be39-ef73bf918cce/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515134142270024444 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015134142271017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015134131660016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015134131661015456 5ustar corecore